Skip to content

Tokenizer & Model info #26

@sbmaruf

Description

@sbmaruf

From the given line, it appears that the data has been transformed using a gpt2 tokenizer.

tokenizer = GPT2TokenizerFast.from_pretrained("gpt2")

However, the model trained is T5.v.1.-lm-adapted checkpoint. Can you confirm that the model was trained with a T5 tokenizer?

Also all the model cards are same in huggingface,

allenai/tk-instruct-11b-def-pos-neg-expl, allenai/tk-instruct-11b-def-pos, allenai/tk-instruct-11b-def.

Can you confirm that the task encoding for these models are following,

  1. allenai/tk-instruct-11b-def-pos-neg-expl : {"add_task_name": False, "add_task_definition": True, "num_pos_examples": 1, "num_neg_examples": 1, "add_explanation": True},
  2. allenai/tk-instruct-11b-def-pos : {"add_task_name": False, "add_task_definition": True, "num_pos_examples": 1, "num_neg_examples": 0, "add_explanation": False},
  3. allenai/tk-instruct-11b-def : {"add_task_name": False, "add_task_definition": True, "num_pos_examples": False, "num_neg_examples": 0, "add_explanation": False},

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions