We propose a simple method for tracking a nonparameterized subject contour in a single video stream with a moving camera. Then we eliminate the tracked contour object by replacing the background scene we get from other frame that is not occluded by...
We propose a simple method for tracking a nonparameterized subject contour in a single video stream with a moving camera. Then we eliminate the tracked contour object by replacing the background scene we get from other frame that is not occluded by the tracked object. Our method consists of two parts: first we track the object using LOD (Level-of-Detail) canny edge maps, then we generate background of each image frame and replace the tracked object in a scene by a background image from other frame. In order to track a contour object, LOD Canny edge maps are generated by changing scale parameters for a given image. A simple (strong) Canny edge map has the smallest number of edge pixels while the most detailed Canny edge map, Wcanny_(N) , has the largest number of edge pixels. To reduce side-effects because of irrelevant edges, we start our basic tracking by using simple (strong) Canny edges generated from large image intensity gradients of an input image, called Scanny edges. Starting from Scanny edges, we get more edge pixels ranging from simple Canny edge maps until the most detailed (weaker) Canny edge maps, called Wcanny maps along LOD hierarchy. LOD Canny edge pixels become nodes in routing, and LOD values of adjacent edge pixels determine routing costs between the nodes. We find the best route to follow Canny edge pixels favoring stronger Canny edge pixels. The first frame background scene is determined by camera motion, and other background scenes are computed from the previous background scenes.