Skip to main content
European Commission logo
English English
CORDIS - EU research results
CORDIS
CORDIS Web 30th anniversary CORDIS Web 30th anniversary

Controlling Large Language Models

Project description

Interpreting and controlling large language models

Large language models (LMs) have rapidly become the backbone of most AI systems, driving cutting-edge advancements across various tasks and applications. However, these benefits come with notable drawbacks, as AI systems often exhibit flaws related to their underlying LMs, such as biased behaviour, confabulations, flawed reasoning and outdated information. These issues have become increasingly difficult to address due to the black-box nature of LMs. The ERC-funded Control-LM project will develop a framework to overcome this opacity, elucidating the internal mechanisms of LMs and enabling safer, more efficient control and interpretation of these models.

Objective

Large language models (LMs) are quickly becoming the backbone of many artificial intelligence (AI) systems, achieving state-of-the-art results in many tasks and application domains. Despite the rapid progress in the field, AI systems suffer from multiple flaws inherited from the underlying LMs: biased behavior, out-of-date information, confabulations, flawed reasoning, and more.
If we wish to control these systems, we must first understand how they work, and develop mechanisms to intervene, update, and repair them. However, the black-box nature of LMs makes them largely inaccessible to such interventions. In this proposal, our overarching goal is to:

*Develop a framework for elucidating the internal mechanisms in LMs and for controlling their behavior in an efficient, interpretable, and safe manner.*

To achieve this goal, we will work through four objectives. First, we will dissect the internal mechanisms of information storage and recall in LMs, and develop ways to update and repair such information.
Second, we will illuminate the mechanisms of higher-level capabilities of LMS to perform reasoning and simulations. We will also repair problems stemming from alignment steps. Third, we will investigate how training processes of LMs affect their emergent mechanisms and develop methods for fine-grained control over the training process. Finally, we will establish a standard benchmark for mechanistic interpretability of LMs to consolidate disparate efforts in the community.
Taken as a whole, we expect the proposed research to empower different stakeholders and ensure a safe, beneficial, and responsible adoption of LMs in AI technologies by our society.

Host institution

TECHNION - ISRAEL INSTITUTE OF TECHNOLOGY
Net EU contribution
€ 1 500 000,00
Address
SENATE BUILDING TECHNION CITY
32000 Haifa
Israel

See on map

Activity type
Higher or Secondary Education Establishments
Links
Total cost
€ 1 500 000,00

Beneficiaries (1)