Omni-Supervised Learning for Dynamic Scene Understanding

This project aims to enhance dynamic scene understanding in autonomous vehicles by developing innovative machine learning models and methods for open-world object recognition from unlabeled video data.

Subsidie
€ 1.500.000
2023

Projectdetails

Introduction

Computer vision has become a powerful technology, able to bring applications such as autonomous vehicles and social robots closer to reality. In order for autonomous vehicles to safely navigate a scene, they need to understand the dynamic objects around them.

Dynamic Scene Understanding

In other words, we need computer vision algorithms to perform dynamic scene understanding (DSU), i.e., detection, segmentation, and tracking of multiple moving objects in a scene. This is an essential feature for higher-level tasks such as action recognition or decision making for autonomous vehicles.

Challenges in Current Models

Much of the success of computer vision models for DSU has been driven by the rise of deep learning, in particular, convolutional neural networks trained on large-scale datasets in a supervised way. However, the closed-world created by our datasets is not an accurate representation of the real world.

If our methods only work on annotated object classes, what happens if a new object appears in front of an autonomous vehicle?

Proposed Solutions

We propose to rethink the deep learning models we use, the way we obtain data annotations, as well as the generalization of our models to previously unseen object classes. To bring all the power of computer vision algorithms for DSU to the open-world, we will focus on three lines of research:

  1. Models: We will design novel machine learning models to address the shortcomings of convolutional neural networks. A hierarchical (from pixels to objects) image-dependent representation will allow us to capture spatio-temporal dependencies at all levels of the hierarchy.

  2. Data: To train our models, we will create a new large-scale DSU synthetic dataset and propose novel methods to mitigate the annotation costs for video data.

  3. Open-World: To bring DSU to the open-world, we will design methods that learn directly from unlabeled video streams. Our models will be able to detect, segment, retrieve, and track dynamic objects coming from classes never previously observed during the training of our models.

Financiële details & Tijdlijn

Financiële details

Subsidiebedrag€ 1.500.000
Totale projectbegroting€ 1.500.000

Tijdlijn

Startdatum1-1-2023
Einddatum31-12-2027
Subsidiejaar2023

Partners & Locaties

Projectpartners

  • NVIDIA ITALY S.R.L.penvoerder

Land(en)

Italy

Vergelijkbare projecten binnen European Research Council

ERC STG

MANUNKIND: Determinants and Dynamics of Collaborative Exploitation

This project aims to develop a game theoretic framework to analyze the psychological and strategic dynamics of collaborative exploitation, informing policies to combat modern slavery.

€ 1.497.749
ERC STG

Elucidating the phenotypic convergence of proliferation reduction under growth-induced pressure

The UnderPressure project aims to investigate how mechanical constraints from 3D crowding affect cell proliferation and signaling in various organisms, with potential applications in reducing cancer chemoresistance.

€ 1.498.280
ERC STG

Uncovering the mechanisms of action of an antiviral bacterium

This project aims to uncover the mechanisms behind Wolbachia's antiviral protection in insects and develop tools for studying symbiont gene function.

€ 1.500.000
ERC STG

The Ethics of Loneliness and Sociability

This project aims to develop a normative theory of loneliness by analyzing ethical responsibilities of individuals and societies to prevent and alleviate loneliness, establishing a new philosophical sub-field.

€ 1.025.860

Vergelijkbare projecten uit andere regelingen

ERC COG

3D scene understanding in two glances

This project aims to understand how adults and infants segment visual scenes by developing new display technologies and using machine learning to simulate and analyze 3D vision inputs.

€ 2.126.444
ERC ADG

Exploration of Unknown Environments for Digital Twins

The 'explorer' project aims to automate video data capture and labeling in open worlds to facilitate the creation of semantically rich Digital Twins for complex environments using AI-driven methods.

€ 2.476.718
ERC ADG

Federated foundational models for embodied perception

The FRONTIER project aims to develop foundational models for embodied perception by integrating neural networks with physical simulations, enhancing learning efficiency and collaboration across intelligent systems.

€ 2.499.825
ERC COG

Learning to synthesize interactive 3D models

This project aims to automate the generation of interactive 3D models using deep learning to enhance virtual environments and applications in animation, robotics, and digital entertainment.

€ 2.000.000