Skip to content

Implement and perform the interpretability analysis of the BigScience models #2

@oserikov

Description

@oserikov

duration: scalable, can be both 175 and 350 hours
mentor: @oserikov
difficulty: easy
requirements:

  1. pytorch
  2. sklearn
  3. experience with re-using the academic code
  4. experience with Transformer Language models

useful links:

Idea Description:

During the season 2021/22, the BigScience team reached several crucial milestones by producing large-scale transformer language models. Some of them even come with the training checkpoints archived, thus allowing to study the emergence of the structures in language models. During this task, we propose to cover the released models with the supplementary interpretability information by applying classical XAI and probing methods described in the attached papers.

Coding Challenge

To better feel what the interpretability work looks like, we ask you to perform a diagnostic classification study of the GPT-like language model, using the SentEval data. Reach out to mentors as soon as possible to discuss the analysis results.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions