Cameras and Algorithms Lab

We conduct research on cognitive and cybernetic vision systems to automate processes. The research activity of the group includes developing models for visual information processing and the development of algorithms that learn in a supervised and self-supervised manner from continuous and limited data as well as from clean and noisy data. Our work is firmly grounded in applications. The lab focuses on transferring the developed technologies to innovative single and multi-camera systems, aided by additional modalities, actuators, and efficient cloud and edge computing.

BP-EVD: Forward Block-Output Propagation for Efficient Video Denoising

We conduct research on cognitive and cybernetic vision systems to automate processes. The research activity of the group includes developing models for visual information processing and the development of algorithms that learn in a supervised and self-supervised manner from continuous and limited data as well as from clean and noisy data. Our work is firmly grounded in applications. The lab focuses on transferring the developed technologies to innovative single and multi-camera systems, aided by additional modalities, actuators, and efficient cloud and edge computing.

Multi-task Video Enhancement for Dental Interventions

A microcamera firmly attached to a dental handpiece allows dentists to continuously monitor the progress of conservative dental procedures. Video enhancement in video-assisted dental interventions alleviates low-light, noise, blur, and camera handshakes that collectively degrade visual comfort. To this end, we introduce a novel deep network for multi-task video enhancement that enables macro-visualization of dental scenes. In particular, the proposed network jointly leverages video restoration and temporal alignment in a multi-scale manner for effective video enhancement.

Concurrent video denoising and deblurring for dynamic scenes

Dynamic scene video deblurring is a challenging task due to the spatially variant blur inflicted by independently moving objects and camera shakes. This is a difficult yet simplified scenario because noise is neglected when it is omnipresent in a wide spectrum of video processing applications. Despite its relevance, the problem of concurrent noise and dynamic blur has not yet been addressed in the deep learning literature. To this end, we analyze existing state-of-the-art deblurring methods and encounter their limitations in handling non-uniform blur under strong noise conditions.

Cameras and Algorithms Lab

Our group conducts basic research on cognitive and cybernetic vision systems to automate processes. We develop models for visual information processing and supervised and self-supervised machine learning algorithms that train the models on temporal, continuous, scarce, clean, and noisy data. Our work is firmly grounded in applications. The lab focuses on transferring the developed technologies to innovative single and multi-camera systems, aided by additional modalities, actuators, and efficient cloud and edge computing.

BP-EVD: Forward Block-Output Propagation for Efficient Video Denoising

Denoising videos in real-time is critical in many applications, including robotics and medicine, where varying-light conditions, miniaturized sensors, and optics can substantially compromise image quality. This work proposes the first video denoising method based on a deep neural network that achieves state-of-the-art performance on dynamic scenes while running in real-time on VGA video resolution with no frame latency. The backbone of our method is a novel, remarkably simple, temporal network of cascaded blocks with forward block output propagation. We train our architecture with short, long, and global residual connections by minimizing the restoration loss of pairs of frames, leading to a more effective training across noise levels. It is robust to heavy noise following Poisson-Gaussian noise statistics. The algorithm is evaluated on RAW and RGB data. We propose a denoising algorithm that requires no future frames to denoise a current frame, reducing its latency considerably. The visual and quantitative results show that our algorithm achieves state-of-the-art performance among efficient algorithms, achieving from two-fold to two-orders-of-magnitude speed-ups on standard benchmarks for video denoising.

Multi-task Video Enhancement for Dental Interventions

A microcamera firmly attached to a dental handpiece allows dentists to continuously monitor the progress of conservative dental procedures. Video enhancement in video-assisted dental interventions alleviates low-light, noise, blur, and camera handshakes that collectively degrade visual comfort. To this end, we introduce a novel deep network for multi-task video enhancement that enables macro-visualization of dental scenes. In particular, the proposed network jointly leverages video restoration and temporal alignment in a multi-scale manner for effective video enhancement.

Concurrent video denoising and deblurring for dynamic scenes

Dynamic scene video deblurring is a challenging task due to the spatially variant blur inflicted by independently moving objects and camera shakes. This is a difficult yet simplified scenario because noise is neglected when it is omnipresent in a wide spectrum of video processing applications. Despite its relevance, the problem of concurrent noise and dynamic blur has not yet been addressed in the deep learning literature. To this end, we analyze existing state-of-the-art deblurring methods and encounter their limitations in handling non-uniform blur under strong noise conditions.