Advanced technologies have recently empowered new smart tiny devices and visual sensors for a wide range of applications including visual surveillance and mobile augmented reality. These advanced devices are able to address the emerging application challenges by cooperatively using multiple visual sensors. They can also enrich visual information by capturing multi-view images and efficiently interact with the environment information within heterogeneous networks in real-time. However, these smart tiny cameras have limited resources for such time-critical applications. In particular, multi-view object tracking and recognition are challenging in a distributed sensing scenario of cooperatively performing tiny cameras. Gathering and processing information can be further enhanced by utilizing correlations among the multi-view images without communications among them.
This project aims at developing novel algorithms and methods by focusing on designing a novel adaptive signal processing framework for multiple distributed visual data sources targeted to visual understanding and mobile augmented reality applications. The project results in developing algorithms and theoretical results, in particular, two of the highlighted applications, rate-efficient multiview object recognition and video separation from compressive measurements as follows:
1. Rate-efficient multiview object recognition: In support of applications involving multiview sources in distributed object recognition using lightweight cameras, we propose a new method for the distributed coding of sparse sources as visual descriptor histograms extracted from multiview images. The problem is challenging due to the computational and energy constraints at each camera as well as the limitations regarding inter camera communication. Our approach addresses these challenges by exploiting the sparsity of the visual descriptor histograms as well as their intra- and inter-camera correlations. Our method couples distributed source coding of the sparse sources with a new joint recovery algorithm that incorporates multiple side information signals, where prior knowledge (low quality) of all the sparse sources is initially sent to exploit their correlations. Experimental evaluation using the histograms of shift-invariant feature transform (SIFT) descriptors extracted from multiview images shows that our method leads to bit-rate saving of up to 43% compared to the state-of-the-art distributed compressed sensing method with the independent encoding of the sources.
2. Video separation from compressive measurements: This work considers online robust principal component analysis (RPCA) in time-varying decomposition problems such as video foreground-background separation. We propose a compressive online RPCA algorithm that decomposes recursively a sequence of data vectors (e.g., frames) into sparse and low-rank components. Different from conventional batch RPCA, which processes all the data directly, our approach considers a small set of measurements taken per data vector (frame). Moreover, our algorithm can incorporate multiple prior information from previous decomposed vectors via proposing an n -ℓ 1 minimization method. At each time instance, the algorithm recovers the sparse vector by solving the n -ℓ 1 minimization problem-which promotes not only the sparsity of the vector but also its correlation with multiple previously recovered sparse vectors-and, subsequently, updates the low-rank component using incremental singular value decomposition. We also establish theoretical bounds on the number of measurements required to guarantee successful compressive separation under the assumptions of static or slowly changing low-rank components. We evaluate the proposed algorithm using numerical experiments and online video foreground-background separation experiments. The experimental results show that the proposed method outperforms the existing methods.
("Compressive Online Robust Principal Component Analysis Via n-l1 Minimization," IEEE Transactions on Image Processing, vol. 27, no. 9, pp. 4314-4329, Sep. 2018.)
References
Copyright notice: Copyright of the above material belongs to the author. The material may not be copied or re-used without permission from the author.