In visual object tracking, occlusions significantly undermine the performance of tracking algorithms. RGB-D cameras, such as Microsoft Kinect or the related PrimeSense camera, are widely available to consumers. Great attention has been focused on exploiting depth information for object tracking in recent years. We propose an algorithm that improves the existing correlation filter-based tracker for scale-adaptive tracking. Moreover, we utilize depth information provided by the Kinect camera to handle various types of occlusions. First, the optimal location of the target is obtained by the conventional kernelized correlation filter tracker. Then, we make use of the discriminative correlation filter for scale estimation as an independent part. At last, to further improve the tracking performance under occlusions, we present a simple yet effective occlusion handling mechanism to detect occlusion and recovery. In this mechanism, cluster analysis and object segmentation by -means method have been applied to depth data. Numerous experiments on Princeton RGB-D tracking dataset demonstrate that the proposed algorithm outperforms several state-of-the-art trackers by successfully dealing with occlusions.