Video foreground segmentation lays the foundation for many high-level visual applications. However, how to dig up the effective features for foreground propagation and how to intelligently fuse the different information are still challenging problems. We aim to deal with the above-mentioned problems, and the goal is to accurately propagate the object across the rest of the frames given an initially labeled frame. Our contributions are summarized as follows: (1) we describe the object features with superpixel-based appearance and motion clues from both global and local viewpoints. Furthermore, the objective confidences for both the appearance and motion features are also introduced to balance the different clues. (2) All the features and their confidences are intelligently fused by the improved Dempster–Shafer evidence theory instead of the empirical parameters tuning used in many algorithms. Experimental results on the two well-known SegTrack and SegTrack v2 datasets demonstrate that our algorithm can yield high-quality segmentations.