Farhand, SepehrTsechpenakis, Gavriil2024-01-082024-01-082023-01Farhand, S., & Tsechpenakis, G. (2023). Foreground discovery in streaming videos with dynamic construction of content graphs. Computer Vision and Image Understanding, 227, 103620. https://doi.org/10.1016/j.cviu.2022.103620https://hdl.handle.net/1805/37713We study the problem of unknown foreground discovery in image streaming scenarios, where no prior information about the dynamic scene is assumed. Contrary to existing co-segmentation principles where the entire dataset is given, in streams new information emerges as content appears and disappears continually. Any object classes to be observed in the scene are unknown, therefore no detection model can be trained for the specific class set. We also assume there is no available repository of trained features from convolutional neural nets, i.e., transfer learning is not applicable. We focus on the progressive discovery of foreground, which may or may not correspond to contextual objects of interest, depending on the camera trajectory, or, in general, the perceived motion. Without any form of supervision, we construct in a bottom up fashion dynamic graphs that capture region saliency and relative topology. Such graphs are continually updated over time, and along with occlusion information, as fundamental property of the foreground–background relationship, foreground is computed for each frame of the stream. We validate our method using indoor and outdoor scenes of varying complexity with respect to content, objects motion, camera trajectory, and occlusions.en-USPublisher Policyimage streamingforegroundregion saliencyrelative topologyForeground discovery in streaming videos with dynamic construction of content graphsArticle