-
Notifications
You must be signed in to change notification settings - Fork 1
Description
duration: scalable, can be both 175 and 350 hours
mentor: @oserikov
difficulty: easy
requirements:
- pytorch
- sklearn
- experience with re-using the academic code
- experience with Transformer Language models
useful links:
- Models produced by BigScience
- BigScience Interpretability papers curated list
- Survey on probing classifiers
- A Primer on Bertology
Idea Description:
During the season 2021/22, the BigScience team reached several crucial milestones by producing large-scale transformer language models. Some of them even come with the training checkpoints archived, thus allowing to study the emergence of the structures in language models. During this task, we propose to cover the released models with the supplementary interpretability information by applying classical XAI and probing methods described in the attached papers.
Coding Challenge
To better feel what the interpretability work looks like, we ask you to perform a diagnostic classification study of the GPT-like language model, using the SentEval data. Reach out to mentors as soon as possible to discuss the analysis results.