The Visionlab is a leader in innovative research within the field of computer vision. Directed by Luigi Cinque, our lab develops groundbreaking technologies and methodologies that significantly advance how computers interpret visual data, enhancing machine-human interactions and driving the evolution of autonomous systems.

Research Areas:

  • Image and Video Analysis:
    Our laboratory specializes in the development of state-of-the-art techniques for image segmentation, motion detection, and visual saliency. These capabilities are crucial for applications ranging from environmental monitoring to advanced video editing and content curation tools.
  • Machine/Deep Learning for Vision:
    We employ advanced statistical models and neural networks to teach machines how to interpret and understand visual information. Our work includes the development of novel training methods that reduce the need for labeled data and the exploration of generative models to create high-fidelity visual simulations for training purposes.
  • Human-Computer Interaction:
    We are pioneering new methods for real-time multi-modal interaction, integrating visual data with audio and tactile feedback to create more immersive and intuitive user experiences. Our research includes the development of advanced virtual assistants that use visual cues to understand user needs better.
  • Autonomous Systems:
    Our laboratory is at the forefront of developing computer vision technologies for autonomous robotics in complex and unstructured environments. This includes deep reinforcement learning models that enable robots to make decisions based on visual inputs and advanced perception systems that can adapt to various lighting and weather conditions.
  • Medical Imaging and Healthcare Applications:
    We explore the application of computer vision in medical imaging, including diagnosis assistance, surgical guidance, and patient monitoring. Our research aims to improve the accuracy and efficiency of medical procedures while reducing healthcare costs and improving patient outcomes.
  • Visual Scene Understanding:
    Our laboratory is dedicated to advancing the understanding of visual scenes, including object detection, scene segmentation, and contextual understanding. This research is fundamental for various applications such as autonomous navigation, augmented reality, and urban planning.
  • Cross-Modal Fusion:
    We investigate methods for integrating information from multiple sensory modalities, such as vision, audio, and text, to improve understanding and decision-making capabilities. This research has applications in multimedia analysis, assistive technologies, and human-robot interaction.
  • Privacy-Preserving Computer Vision:
    In response to growing concerns about privacy and data security, our lab develops techniques for performing computer vision tasks while preserving the privacy of individuals’ visual data. This includes methods for anonymization, secure computation, and decentralized processing.

 

Video Understanding

Video understanding stands as a critical pillar of AI research, encompassing the multifaceted challenge of interpreting the rich visual and temporal information embedded in videos. Within this field, two key branches emerge: egocentric video understanding and exocentric video understanding, each posing unique challenges and opportunities.

Egocentric video understanding involves interpreting footage captured from a first-person perspective, offering a subjective view of the world as experienced by the wearer of the camera. This perspective provides invaluable insights into human activities, intentions, and interactions with the environment, enabling applications in areas like augmented reality, assistive technologies, and even psychological studies.

The alignment of egocentric and exocentric videos presents a fascinating avenue for research, aiming to bridge the gap between these two contrasting perspectives. By establishing correspondence between events, objects, and actions as observed from both viewpoints, AI systems can gain a deeper understanding of the world, enabling more sophisticated and context-aware applications. This alignment also holds the key to unlocking new possibilities for cross-view learning and knowledge transfer, where insights gained from one perspective can enrich the understanding of the other.

Our research in this field involves national and international collaborations with the University of Catania, the Technical University of Munich, and big tech companies such as Huawei.

Associated Researchers:

– Marco Raoul Marini

– Romeo Lanzino

– Federico Fontana

– Anxhelo Diko

– Alessio Fagioli