Simon Ostermann

DFKI

Software Project: Recent Advances in Mechanistic Interpretability

time/location TBD

Tanja Bäumel, Tatiana Anikina, Dr. Simon Ostermann
Deutsches Forschungszentrum für Künstliche Intelligenz (DFKI)

Prerequisites

This seminar is targeted at Master students and will require students to implement mechanistic approaches on a low level of language models. We thus expect you to have a curious mind and advanced familiarity with large language models. If you would like to participate, you need to visit the Mechanistic Interpretability seminar or proof to us otherwise that you bring the necessary knowledge to implement mechanistic approaches.

This is a software project that will take place during the semester break after the lecture period in the next winter semester, i.e. in February and March 2025. If you want to take part, please drop an email to mechanistic-interpretability-seminar@googlegroups.com (deadline TBD). In your email, please:

Project Timeline

Project Content

The rise of deep learning in AI has dramatically increased the performance of models across many sub-fields such as natural language processing or computer vision. In the last 5 years, large pretrained language models (LLMs) and their variants (BERT, ChatGPT etc.) have changed the NLP landscape drastically. Such models got larger and larger over the last years, reaching increasingly impressive performance peeks, sometimes even surpassing humans.

A central issue with deep learning models with millions or billions of parameters is that they are essentially black boxes: From the model's parameters, it is not inherently clear why a model exhibits a certain behavior or makes certain classification decision. The rapidly growing field of interpretable and explainable AI (XAI) develops methods to peek into the black box that LLMs are, trying to understand the inner workings of such large models.

In this seminar we will investiagte a subfield in XAI, namely Mechanistic Interpretability (MI). MI aims to understand and explain the internal workings of complex machine learning models, in particular deep neural networks, by "reverse-engineering" internal mechanisms and computations inside a network's parameters. This involves analysing how certain mechanisms in the model contribute to decisions and results. The aim is to break down the ‘black box’ nature of these models and reveal the underlying calculations and internal representations that lead to predictions or behaviours.

Students in the course will be introduced to a range of toolkits and methods inside the field of MI and will pick their own topic (single students or groups of two) within MI. They will invetigate a language model of their choice with a technique of their choice for some specific phenomenom and submit both code for reproducing their experiments and a project report.

Projects can (but don't have to) be based on Neel Nanda's list of 200 open problems in mechanistic interpretability.

Back to Main Page