DEep COgnition Learning for LAnguage GEneration

This project aims to enhance NLP models by integrating machine learning, cognitive science, and structured memory to improve out-of-domain generalization and contextual understanding in language generation tasks.

Subsidie
€ 1.999.595
2023

Projectdetails

Introduction

In recent years, transformer-based deep learning models such as BERT or GPT-3 have led to impressive results in many natural language processing (NLP) tasks, exhibiting transfer and few-shot learning capabilities.

Limitations of Current Models

However, despite faring well in benchmarks, current deep learning models for NLP often fail badly in the wild. The main issues include:

  • Poor out-of-domain generalization
  • Inability to exploit contextual information
  • Poor calibration
  • Non-traceable memory

These limitations stem from their monolithic architectures, which are good for perception but unsuitable for tasks requiring higher-level cognition.

Project Goals

In this project, I attack these fundamental problems by bringing together tools and ideas from various fields, including:

  1. Machine learning
  2. Sparse modeling
  3. Information theory
  4. Cognitive science

This interdisciplinary approach aims to address the limitations of current models.

Methodology

Utility-Guided Controlled Generation

First, I will use uncertainty and quality estimates for utility-guided controlled generation. This will involve:

  • Combining the control mechanism with efficient encoding of contextual information
  • Integrating multiple modalities

Sparse and Structured Memory Models

Second, I will develop sparse and structured memory models, along with attention descriptive representations aimed at conscious processing.

Mathematical Models for Sparse Communication

Third, I will build mathematical models for sparse communication, which will involve:

  • Reconciling discrete and continuous domains
  • Supporting end-to-end differentiability
  • Enabling a shared workspace for communication among multiple modules and agents

Application of Innovations

I will apply the innovations above to highly challenging language generation tasks, including:

  • Machine translation
  • Open dialogue
  • Story generation

Collaborations

To reinforce interdisciplinarity and maximize technological impact, collaborations are planned with:

  • Cognitive scientists
  • A scale-up company in the crowd-sourcing translation industry

Financiële details & Tijdlijn

Financiële details

Subsidiebedrag€ 1.999.595
Totale projectbegroting€ 1.999.595

Tijdlijn

Startdatum1-8-2023
Einddatum31-7-2028
Subsidiejaar2023

Partners & Locaties

Projectpartners

  • INSTITUTO DE TELECOMUNICACOESpenvoerder
  • UNBABEL UNIPESSOAL, LDA

Land(en)

Portugal

Vergelijkbare projecten binnen European Research Council

ERC STG

MANUNKIND: Determinants and Dynamics of Collaborative Exploitation

This project aims to develop a game theoretic framework to analyze the psychological and strategic dynamics of collaborative exploitation, informing policies to combat modern slavery.

€ 1.497.749
ERC STG

Elucidating the phenotypic convergence of proliferation reduction under growth-induced pressure

The UnderPressure project aims to investigate how mechanical constraints from 3D crowding affect cell proliferation and signaling in various organisms, with potential applications in reducing cancer chemoresistance.

€ 1.498.280
ERC STG

Uncovering the mechanisms of action of an antiviral bacterium

This project aims to uncover the mechanisms behind Wolbachia's antiviral protection in insects and develop tools for studying symbiont gene function.

€ 1.500.000
ERC STG

The Ethics of Loneliness and Sociability

This project aims to develop a normative theory of loneliness by analyzing ethical responsibilities of individuals and societies to prevent and alleviate loneliness, establishing a new philosophical sub-field.

€ 1.025.860

Vergelijkbare projecten uit andere regelingen

ERC STG

Next-Generation Natural Language Generation

This project aims to enhance natural language generation by integrating neural models with symbolic representations for better control, adaptability, and reliable evaluation across various applications.

€ 1.420.375
ERC STG

Towards an Artificial Cognitive Science

This project aims to establish a new field of artificial cognitive science by applying cognitive psychology to enhance the learning and decision-making of advanced AI models.

€ 1.496.000
ERC ADG

Why do infants learn language so fast? A reverse engineering approach

This project develops a computational model to explore how infants efficiently learn language through statistical learning and three additional mechanisms, aiming to produce comparable outcomes to children's language acquisition.

€ 2.494.625
ERC ADG

Evaluating and Programming Intelligent Chatbots for Any Language

EPICAL aims to enhance intelligent chatbots by integrating low resource languages through innovative techniques in text generation, translation, and speech processing, promoting multilingual inclusivity.

€ 2.498.200