diff --git a/src/lighteval/tasks/default_tasks.py b/src/lighteval/tasks/default_tasks.py index 7abecc929..7e3a888b4 100644 --- a/src/lighteval/tasks/default_tasks.py +++ b/src/lighteval/tasks/default_tasks.py @@ -79,8 +79,8 @@ version=0, ) abstract_narrative_understanding_bigbench = LightevalTaskConfig( - name="abstract_narrative_understanding", - suite=["bigbench", "bigbench_json"], + name="bigbench:abstract_narrative_understanding", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="abstract_narrative_understanding", @@ -470,8 +470,8 @@ version=1, ) anachronisms_bigbench = LightevalTaskConfig( - name="anachronisms", - suite=["bigbench", "bigbench_json"], + name="bigbench:anachronisms", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="anachronisms", @@ -485,8 +485,8 @@ version=0, ) analogical_similarity_bigbench = LightevalTaskConfig( - name="analogical_similarity", - suite=["bigbench", "bigbench_json"], + name="bigbench:analogical_similarity", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="analogical_similarity", @@ -500,8 +500,8 @@ version=0, ) analytic_entailment_bigbench = LightevalTaskConfig( - name="analytic_entailment", - suite=["bigbench", "bigbench_json"], + name="bigbench:analytic_entailment", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="analytic_entailment", @@ -812,8 +812,8 @@ version=0, ) arithmetic_bb_bigbench = LightevalTaskConfig( - name="arithmetic_bb", - suite=["bigbench", "bigbench_json"], + name="bigbench:arithmetic_bb", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="arithmetic", @@ -827,8 +827,8 @@ version=0, ) ascii_word_recognition_bigbench = LightevalTaskConfig( - name="ascii_word_recognition", - suite=["bigbench", "bigbench_json"], + name="bigbench:ascii_word_recognition", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="ascii_word_recognition", @@ -857,8 +857,8 @@ version=0, ) authorship_verification_bigbench = LightevalTaskConfig( - name="authorship_verification", - suite=["bigbench", "bigbench_json"], + name="bigbench:authorship_verification", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="authorship_verification", @@ -872,8 +872,8 @@ version=0, ) auto_categorization_bigbench = LightevalTaskConfig( - name="auto_categorization", - suite=["bigbench", "bigbench_json"], + name="bigbench:auto_categorization", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="auto_categorization", @@ -887,8 +887,8 @@ version=0, ) auto_debugging_bigbench_lite = LightevalTaskConfig( - name="auto_debugging", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:auto_debugging", + suite=["lighteval"], prompt_function=prompt.bigbench_linefeed_before_and_after_query, hf_repo="tasksource/bigbench", hf_subset="auto_debugging", @@ -2593,8 +2593,8 @@ version=0, ) bbq_lite_json_bigbench_lite = LightevalTaskConfig( - name="bbq_lite_json", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:bbq_lite_json", + suite=["lighteval"], prompt_function=prompt.bigbench_linefeed_before_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="bbq_lite_json", @@ -6487,8 +6487,8 @@ version=0, ) bridging_anaphora_resolution_barqa_bigbench = LightevalTaskConfig( - name="bridging_anaphora_resolution_barqa", - suite=["bigbench", "bigbench_json"], + name="bigbench:bridging_anaphora_resolution_barqa", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="bridging_anaphora_resolution_barqa", @@ -6502,8 +6502,8 @@ version=0, ) causal_judgment_bigbench = LightevalTaskConfig( - name="causal_judgment", - suite=["bigbench", "bigbench_json"], + name="bigbench:causal_judgment", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="causal_judgment", @@ -6517,8 +6517,8 @@ version=0, ) cause_and_effect_bigbench = LightevalTaskConfig( - name="cause_and_effect", - suite=["bigbench", "bigbench_json"], + name="bigbench:cause_and_effect", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="cause_and_effect", @@ -6532,8 +6532,8 @@ version=0, ) checkmate_in_one_bigbench = LightevalTaskConfig( - name="checkmate_in_one", - suite=["bigbench", "bigbench_json"], + name="bigbench:checkmate_in_one", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="checkmate_in_one", @@ -6547,8 +6547,8 @@ version=0, ) chess_state_tracking_bigbench = LightevalTaskConfig( - name="chess_state_tracking", - suite=["bigbench", "bigbench_json"], + name="bigbench:chess_state_tracking", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="chess_state_tracking", @@ -6562,8 +6562,8 @@ version=0, ) chinese_remainder_theorem_bigbench = LightevalTaskConfig( - name="chinese_remainder_theorem", - suite=["bigbench", "bigbench_json"], + name="bigbench:chinese_remainder_theorem", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="chinese_remainder_theorem", @@ -6577,8 +6577,8 @@ version=0, ) cifar10_classification_bigbench = LightevalTaskConfig( - name="cifar10_classification", - suite=["bigbench", "bigbench_json"], + name="bigbench:cifar10_classification", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="cifar10_classification", @@ -6844,8 +6844,8 @@ version=0, ) code_line_description_bigbench_lite = LightevalTaskConfig( - name="code_line_description", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:code_line_description", + suite=["lighteval"], prompt_function=prompt.bigbench_linefeed_before_and_after_query, hf_repo="tasksource/bigbench", hf_subset="code_line_description", @@ -6859,8 +6859,8 @@ version=0, ) codenames_bigbench = LightevalTaskConfig( - name="codenames", - suite=["bigbench", "bigbench_json"], + name="bigbench:codenames", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="codenames", @@ -6874,8 +6874,8 @@ version=0, ) color_bigbench = LightevalTaskConfig( - name="color", - suite=["bigbench", "bigbench_json"], + name="bigbench:color", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="color", @@ -6894,8 +6894,8 @@ version=0, ) common_morpheme_bigbench = LightevalTaskConfig( - name="common_morpheme", - suite=["bigbench", "bigbench_json"], + name="bigbench:common_morpheme", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="common_morpheme", @@ -6935,8 +6935,8 @@ version=0, ) conceptual_combinations_bigbench_lite = LightevalTaskConfig( - name="conceptual_combinations", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:conceptual_combinations", + suite=["lighteval"], prompt_function=prompt.bigbench_linefeed_before_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="conceptual_combinations", @@ -6950,8 +6950,8 @@ version=0, ) conlang_translation_bigbench_lite = LightevalTaskConfig( - name="conlang_translation", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:conlang_translation", + suite=["lighteval"], prompt_function=prompt.bigbench_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="conlang_translation", @@ -6965,8 +6965,8 @@ version=0, ) contextual_parametric_knowledge_conflicts_bigbench = LightevalTaskConfig( - name="contextual_parametric_knowledge_conflicts", - suite=["bigbench", "bigbench_json"], + name="bigbench:contextual_parametric_knowledge_conflicts", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="contextual_parametric_knowledge_conflicts", @@ -7301,8 +7301,8 @@ version=0, ) crash_blossom_bigbench = LightevalTaskConfig( - name="crash_blossom", - suite=["bigbench", "bigbench_json"], + name="bigbench:crash_blossom", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="crash_blossom", @@ -7316,8 +7316,8 @@ version=0, ) crass_ai_bigbench = LightevalTaskConfig( - name="crass_ai", - suite=["bigbench", "bigbench_json"], + name="bigbench:crass_ai", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="crass_ai", @@ -7331,8 +7331,8 @@ version=0, ) cryobiology_spanish_bigbench = LightevalTaskConfig( - name="cryobiology_spanish", - suite=["bigbench", "bigbench_json"], + name="bigbench:cryobiology_spanish", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="cryobiology_spanish", @@ -7346,8 +7346,8 @@ version=0, ) cryptonite_bigbench = LightevalTaskConfig( - name="cryptonite", - suite=["bigbench", "bigbench_json"], + name="bigbench:cryptonite", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="cryptonite", @@ -7361,8 +7361,8 @@ version=0, ) cs_algorithms_bigbench = LightevalTaskConfig( - name="cs_algorithms", - suite=["bigbench", "bigbench_json"], + name="bigbench:cs_algorithms", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="cs_algorithms", @@ -7376,8 +7376,8 @@ version=0, ) dark_humor_detection_bigbench = LightevalTaskConfig( - name="dark_humor_detection", - suite=["bigbench", "bigbench_json"], + name="bigbench:dark_humor_detection", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="dark_humor_detection", @@ -7391,8 +7391,8 @@ version=0, ) date_understanding_bigbench = LightevalTaskConfig( - name="date_understanding", - suite=["bigbench", "bigbench_json"], + name="bigbench:date_understanding", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="date_understanding", @@ -7406,8 +7406,8 @@ version=0, ) disambiguation_qa_bigbench = LightevalTaskConfig( - name="disambiguation_qa", - suite=["bigbench", "bigbench_json"], + name="bigbench:disambiguation_qa", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="disambiguation_qa", @@ -7421,8 +7421,8 @@ version=0, ) discourse_marker_prediction_bigbench = LightevalTaskConfig( - name="discourse_marker_prediction", - suite=["bigbench", "bigbench_json"], + name="bigbench:discourse_marker_prediction", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="discourse_marker_prediction", @@ -7436,8 +7436,8 @@ version=0, ) disfl_qa_bigbench = LightevalTaskConfig( - name="disfl_qa", - suite=["bigbench", "bigbench_json"], + name="bigbench:disfl_qa", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="disfl_qa", @@ -7540,8 +7540,8 @@ version=0, ) dyck_languages_bigbench = LightevalTaskConfig( - name="dyck_languages", - suite=["bigbench", "bigbench_json"], + name="bigbench:dyck_languages", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="dyck_languages", @@ -7555,8 +7555,8 @@ version=0, ) elementary_math_qa_bigbench = LightevalTaskConfig( - name="elementary_math_qa", - suite=["bigbench", "bigbench_json"], + name="bigbench:elementary_math_qa", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="elementary_math_qa", @@ -7570,8 +7570,8 @@ version=0, ) emoji_movie_bigbench_lite = LightevalTaskConfig( - name="emoji_movie", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:emoji_movie", + suite=["lighteval"], prompt_function=prompt.bigbench_linefeed_before_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="emoji_movie", @@ -7590,8 +7590,8 @@ version=0, ) emojis_emotion_prediction_bigbench = LightevalTaskConfig( - name="emojis_emotion_prediction", - suite=["bigbench", "bigbench_json"], + name="bigbench:emojis_emotion_prediction", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="emojis_emotion_prediction", @@ -7605,8 +7605,8 @@ version=0, ) empirical_judgments_bigbench = LightevalTaskConfig( - name="empirical_judgments", - suite=["bigbench", "bigbench_json"], + name="bigbench:empirical_judgments", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="empirical_judgments", @@ -7620,8 +7620,8 @@ version=0, ) english_proverbs_bigbench = LightevalTaskConfig( - name="english_proverbs", - suite=["bigbench", "bigbench_json"], + name="bigbench:english_proverbs", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="english_proverbs", @@ -7635,8 +7635,8 @@ version=0, ) english_russian_proverbs_bigbench = LightevalTaskConfig( - name="english_russian_proverbs", - suite=["bigbench", "bigbench_json"], + name="bigbench:english_russian_proverbs", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="english_russian_proverbs", @@ -7650,8 +7650,8 @@ version=0, ) entailed_polarity_bigbench = LightevalTaskConfig( - name="entailed_polarity", - suite=["bigbench", "bigbench_json"], + name="bigbench:entailed_polarity", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="entailed_polarity", @@ -7665,8 +7665,8 @@ version=0, ) entailed_polarity_hindi_bigbench = LightevalTaskConfig( - name="entailed_polarity_hindi", - suite=["bigbench", "bigbench_json"], + name="bigbench:entailed_polarity_hindi", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="entailed_polarity_hindi", @@ -8070,8 +8070,8 @@ version=0, ) epistemic_reasoning_bigbench = LightevalTaskConfig( - name="epistemic_reasoning", - suite=["bigbench", "bigbench_json"], + name="bigbench:epistemic_reasoning", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="epistemic_reasoning", @@ -8160,8 +8160,8 @@ version=0, ) evaluating_information_essentiality_bigbench = LightevalTaskConfig( - name="evaluating_information_essentiality", - suite=["bigbench", "bigbench_json"], + name="bigbench:evaluating_information_essentiality", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="evaluating_information_essentiality", @@ -8175,8 +8175,8 @@ version=0, ) fact_checker_bigbench = LightevalTaskConfig( - name="fact_checker", - suite=["bigbench", "bigbench_json"], + name="bigbench:fact_checker", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="fact_checker", @@ -8190,8 +8190,8 @@ version=0, ) fantasy_reasoning_bigbench = LightevalTaskConfig( - name="fantasy_reasoning", - suite=["bigbench", "bigbench_json"], + name="bigbench:fantasy_reasoning", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="fantasy_reasoning", @@ -8205,8 +8205,8 @@ version=0, ) few_shot_nlg_bigbench = LightevalTaskConfig( - name="few_shot_nlg", - suite=["bigbench", "bigbench_json"], + name="bigbench:few_shot_nlg", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="few_shot_nlg", @@ -8220,8 +8220,8 @@ version=0, ) figure_of_speech_detection_bigbench = LightevalTaskConfig( - name="figure_of_speech_detection", - suite=["bigbench", "bigbench_json"], + name="bigbench:figure_of_speech_detection", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="figure_of_speech_detection", @@ -8235,8 +8235,8 @@ version=0, ) formal_fallacies_syllogisms_negation_bigbench_lite = LightevalTaskConfig( - name="formal_fallacies_syllogisms_negation", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:formal_fallacies_syllogisms_negation", + suite=["lighteval"], prompt_function=prompt.bigbench_linefeed_before_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="formal_fallacies_syllogisms_negation", @@ -8250,8 +8250,8 @@ version=0, ) gem_bigbench = LightevalTaskConfig( - name="gem", - suite=["bigbench", "bigbench_json"], + name="bigbench:gem", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="gem", @@ -8265,8 +8265,8 @@ version=0, ) gender_inclusive_sentences_german_bigbench = LightevalTaskConfig( - name="gender_inclusive_sentences_german", - suite=["bigbench", "bigbench_json"], + name="bigbench:gender_inclusive_sentences_german", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="gender_inclusive_sentences_german", @@ -8280,8 +8280,8 @@ version=0, ) general_knowledge_bigbench = LightevalTaskConfig( - name="general_knowledge", - suite=["bigbench", "bigbench_json"], + name="bigbench:general_knowledge", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="general_knowledge", @@ -8295,8 +8295,8 @@ version=0, ) geometric_shapes_bigbench = LightevalTaskConfig( - name="geometric_shapes", - suite=["bigbench", "bigbench_json"], + name="bigbench:geometric_shapes", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="geometric_shapes", @@ -8465,8 +8465,8 @@ version=0, ) goal_step_wikihow_bigbench = LightevalTaskConfig( - name="goal_step_wikihow", - suite=["bigbench", "bigbench_json"], + name="bigbench:goal_step_wikihow", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="goal_step_wikihow", @@ -8540,8 +8540,8 @@ version=0, ) gre_reading_comprehension_bigbench = LightevalTaskConfig( - name="gre_reading_comprehension", - suite=["bigbench", "bigbench_json"], + name="bigbench:gre_reading_comprehension", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="gre_reading_comprehension", @@ -8684,8 +8684,8 @@ version=0, ) hhh_alignment_bigbench = LightevalTaskConfig( - name="hhh_alignment", - suite=["bigbench", "bigbench_json"], + name="bigbench:hhh_alignment", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="hhh_alignment", @@ -8699,8 +8699,8 @@ version=0, ) hindi_question_answering_bigbench = LightevalTaskConfig( - name="hindi_question_answering", - suite=["bigbench", "bigbench_json"], + name="bigbench:hindi_question_answering", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="hindi_question_answering", @@ -8714,8 +8714,8 @@ version=0, ) hindu_knowledge_bigbench_lite = LightevalTaskConfig( - name="hindu_knowledge", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:hindu_knowledge", + suite=["lighteval"], prompt_function=prompt.bigbench_linefeed_before_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="hindu_knowledge", @@ -8729,8 +8729,8 @@ version=0, ) hinglish_toxicity_bigbench = LightevalTaskConfig( - name="hinglish_toxicity", - suite=["bigbench", "bigbench_json"], + name="bigbench:hinglish_toxicity", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="hinglish_toxicity", @@ -8744,8 +8744,8 @@ version=0, ) human_organs_senses_bigbench = LightevalTaskConfig( - name="human_organs_senses", - suite=["bigbench", "bigbench_json"], + name="bigbench:human_organs_senses", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="human_organs_senses", @@ -8759,8 +8759,8 @@ version=0, ) hyperbaton_bigbench = LightevalTaskConfig( - name="hyperbaton", - suite=["bigbench", "bigbench_json"], + name="bigbench:hyperbaton", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="hyperbaton", @@ -8774,8 +8774,8 @@ version=0, ) identify_math_theorems_bigbench = LightevalTaskConfig( - name="identify_math_theorems", - suite=["bigbench", "bigbench_json"], + name="bigbench:identify_math_theorems", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="identify_math_theorems", @@ -8789,8 +8789,8 @@ version=0, ) identify_odd_metaphor_bigbench = LightevalTaskConfig( - name="identify_odd_metaphor", - suite=["bigbench", "bigbench_json"], + name="bigbench:identify_odd_metaphor", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="identify_odd_metaphor", @@ -8860,8 +8860,8 @@ version=0, ) implicatures_bigbench = LightevalTaskConfig( - name="implicatures", - suite=["bigbench", "bigbench_json"], + name="bigbench:implicatures", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="implicatures", @@ -8875,8 +8875,8 @@ version=0, ) implicit_relations_bigbench = LightevalTaskConfig( - name="implicit_relations", - suite=["bigbench", "bigbench_json"], + name="bigbench:implicit_relations", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="implicit_relations", @@ -8890,8 +8890,8 @@ version=0, ) intent_recognition_bigbench = LightevalTaskConfig( - name="intent_recognition", - suite=["bigbench", "bigbench_json"], + name="bigbench:intent_recognition", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="intent_recognition", @@ -9061,8 +9061,8 @@ version=0, ) international_phonetic_alphabet_nli_bigbench = LightevalTaskConfig( - name="international_phonetic_alphabet_nli", - suite=["bigbench", "bigbench_json"], + name="bigbench:international_phonetic_alphabet_nli", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="international_phonetic_alphabet_nli", @@ -9076,8 +9076,8 @@ version=0, ) international_phonetic_alphabet_transliterate_bigbench = LightevalTaskConfig( - name="international_phonetic_alphabet_transliterate", - suite=["bigbench", "bigbench_json"], + name="bigbench:international_phonetic_alphabet_transliterate", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="international_phonetic_alphabet_transliterate", @@ -9091,8 +9091,8 @@ version=0, ) intersect_geometry_bigbench = LightevalTaskConfig( - name="intersect_geometry", - suite=["bigbench", "bigbench_json"], + name="bigbench:intersect_geometry", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="intersect_geometry", @@ -9106,8 +9106,8 @@ version=0, ) irony_identification_bigbench = LightevalTaskConfig( - name="irony_identification", - suite=["bigbench", "bigbench_json"], + name="bigbench:irony_identification", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="irony_identification", @@ -9328,8 +9328,8 @@ ), ) kanji_ascii_bigbench = LightevalTaskConfig( - name="kanji_ascii", - suite=["bigbench", "bigbench_json"], + name="bigbench:kanji_ascii", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="kanji_ascii", @@ -9343,8 +9343,8 @@ version=0, ) kannada_bigbench = LightevalTaskConfig( - name="kannada", - suite=["bigbench", "bigbench_json"], + name="bigbench:kannada", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="kannada", @@ -9358,8 +9358,8 @@ version=0, ) key_value_maps_bigbench = LightevalTaskConfig( - name="key_value_maps", - suite=["bigbench", "bigbench_json"], + name="bigbench:key_value_maps", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="key_value_maps", @@ -9373,8 +9373,8 @@ version=0, ) known_unknowns_bigbench_lite = LightevalTaskConfig( - name="known_unknowns", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:known_unknowns", + suite=["lighteval"], prompt_function=prompt.bigbench_linefeed_before_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="known_unknowns", @@ -9523,8 +9523,8 @@ version=0, ) language_games_bigbench = LightevalTaskConfig( - name="language_games", - suite=["bigbench", "bigbench_json"], + name="bigbench:language_games", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="language_games", @@ -9538,8 +9538,8 @@ version=0, ) language_identification_bigbench_lite = LightevalTaskConfig( - name="language_identification", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:language_identification", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="language_identification", @@ -10169,8 +10169,8 @@ version=0, ) linguistic_mappings_bigbench = LightevalTaskConfig( - name="linguistic_mappings", - suite=["bigbench", "bigbench_json"], + name="bigbench:linguistic_mappings", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="linguistic_mappings", @@ -10184,8 +10184,8 @@ version=0, ) linguistics_puzzles_bigbench_lite = LightevalTaskConfig( - name="linguistics_puzzles", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:linguistics_puzzles", + suite=["lighteval"], prompt_function=prompt.bigbench_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="linguistics_puzzles", @@ -10199,8 +10199,8 @@ version=0, ) logic_grid_puzzle_bigbench_lite = LightevalTaskConfig( - name="logic_grid_puzzle", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:logic_grid_puzzle", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="logic_grid_puzzle", @@ -10214,8 +10214,8 @@ version=0, ) logical_args_bigbench = LightevalTaskConfig( - name="logical_args", - suite=["bigbench", "bigbench_json"], + name="bigbench:logical_args", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="logical_args", @@ -10229,8 +10229,8 @@ version=0, ) logical_deduction_bigbench_lite = LightevalTaskConfig( - name="logical_deduction", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:logical_deduction", + suite=["lighteval"], prompt_function=prompt.bigbench_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="logical_deduction", @@ -10244,8 +10244,8 @@ version=0, ) logical_fallacy_detection_bigbench = LightevalTaskConfig( - name="logical_fallacy_detection", - suite=["bigbench", "bigbench_json"], + name="bigbench:logical_fallacy_detection", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="logical_fallacy_detection", @@ -10259,8 +10259,8 @@ version=0, ) logical_sequence_bigbench = LightevalTaskConfig( - name="logical_sequence", - suite=["bigbench", "bigbench_json"], + name="bigbench:logical_sequence", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="logical_sequence", @@ -10802,8 +10802,8 @@ version=0, ) mathematical_induction_bigbench = LightevalTaskConfig( - name="mathematical_induction", - suite=["bigbench", "bigbench_json"], + name="bigbench:mathematical_induction", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="mathematical_induction", @@ -10835,8 +10835,8 @@ version=0, ) matrixshapes_bigbench = LightevalTaskConfig( - name="matrixshapes", - suite=["bigbench", "bigbench_json"], + name="bigbench:matrixshapes", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="matrixshapes", @@ -10992,8 +10992,8 @@ version=0, ) metaphor_boolean_bigbench = LightevalTaskConfig( - name="metaphor_boolean", - suite=["bigbench", "bigbench_json"], + name="bigbench:metaphor_boolean", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="metaphor_boolean", @@ -11007,8 +11007,8 @@ version=0, ) metaphor_understanding_bigbench = LightevalTaskConfig( - name="metaphor_understanding", - suite=["bigbench", "bigbench_json"], + name="bigbench:metaphor_understanding", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="metaphor_understanding", @@ -11220,8 +11220,8 @@ version=0, ) minute_mysteries_qa_bigbench = LightevalTaskConfig( - name="minute_mysteries_qa", - suite=["bigbench", "bigbench_json"], + name="bigbench:minute_mysteries_qa", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="minute_mysteries_qa", @@ -11235,8 +11235,8 @@ version=0, ) misconceptions_bigbench = LightevalTaskConfig( - name="misconceptions", - suite=["bigbench", "bigbench_json"], + name="bigbench:misconceptions", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="misconceptions", @@ -11250,8 +11250,8 @@ version=0, ) misconceptions_russian_bigbench_lite = LightevalTaskConfig( - name="misconceptions_russian", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:misconceptions_russian", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="misconceptions_russian", @@ -14457,8 +14457,8 @@ version=0, ) mnist_ascii_bigbench = LightevalTaskConfig( - name="mnist_ascii", - suite=["bigbench", "bigbench_json"], + name="bigbench:mnist_ascii", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="mnist_ascii", @@ -14472,8 +14472,8 @@ version=0, ) modified_arithmetic_bigbench = LightevalTaskConfig( - name="modified_arithmetic", - suite=["bigbench", "bigbench_json"], + name="bigbench:modified_arithmetic", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="modified_arithmetic", @@ -14487,8 +14487,8 @@ version=0, ) moral_permissibility_bigbench = LightevalTaskConfig( - name="moral_permissibility", - suite=["bigbench", "bigbench_json"], + name="bigbench:moral_permissibility", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="moral_permissibility", @@ -14502,8 +14502,8 @@ version=0, ) movie_dialog_same_or_different_bigbench = LightevalTaskConfig( - name="movie_dialog_same_or_different", - suite=["bigbench", "bigbench_json"], + name="bigbench:movie_dialog_same_or_different", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="movie_dialog_same_or_different", @@ -14517,8 +14517,8 @@ version=0, ) movie_recommendation_bigbench = LightevalTaskConfig( - name="movie_recommendation", - suite=["bigbench", "bigbench_json"], + name="bigbench:movie_recommendation", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="movie_recommendation", @@ -14592,8 +14592,8 @@ version=0, ) mult_data_wrangling_bigbench = LightevalTaskConfig( - name="mult_data_wrangling", - suite=["bigbench", "bigbench_json"], + name="bigbench:mult_data_wrangling", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="mult_data_wrangling", @@ -14607,8 +14607,8 @@ version=0, ) multiemo_bigbench = LightevalTaskConfig( - name="multiemo", - suite=["bigbench", "bigbench_json"], + name="bigbench:multiemo", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="multiemo", @@ -14719,8 +14719,8 @@ version=0, ) natural_instructions_bigbench = LightevalTaskConfig( - name="natural_instructions", - suite=["bigbench", "bigbench_json"], + name="bigbench:natural_instructions", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="natural_instructions", @@ -14758,8 +14758,8 @@ ), ) navigate_bigbench = LightevalTaskConfig( - name="navigate", - suite=["bigbench", "bigbench_json"], + name="bigbench:navigate", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="navigate", @@ -14773,8 +14773,8 @@ version=0, ) nonsense_words_grammar_bigbench = LightevalTaskConfig( - name="nonsense_words_grammar", - suite=["bigbench", "bigbench_json"], + name="bigbench:nonsense_words_grammar", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="nonsense_words_grammar", @@ -14788,8 +14788,8 @@ version=0, ) novel_concepts_bigbench_lite = LightevalTaskConfig( - name="novel_concepts", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:novel_concepts", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="novel_concepts", @@ -14947,8 +14947,8 @@ version=0, ) object_counting_bigbench = LightevalTaskConfig( - name="object_counting", - suite=["bigbench", "bigbench_json"], + name="bigbench:object_counting", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="object_counting", @@ -14962,8 +14962,8 @@ version=0, ) odd_one_out_bigbench = LightevalTaskConfig( - name="odd_one_out", - suite=["bigbench", "bigbench_json"], + name="bigbench:odd_one_out", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="odd_one_out", @@ -15021,8 +15021,8 @@ version=0, ) operators_bigbench_lite = LightevalTaskConfig( - name="operators", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:operators", + suite=["lighteval"], prompt_function=prompt.bigbench_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="operators", @@ -15036,8 +15036,8 @@ version=0, ) paragraph_segmentation_bigbench = LightevalTaskConfig( - name="paragraph_segmentation", - suite=["bigbench", "bigbench_json"], + name="bigbench:paragraph_segmentation", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="paragraph_segmentation", @@ -15051,8 +15051,8 @@ version=0, ) parsinlu_qa_bigbench = LightevalTaskConfig( - name="parsinlu_qa", - suite=["bigbench", "bigbench_json"], + name="bigbench:parsinlu_qa", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="parsinlu_qa", @@ -15066,8 +15066,8 @@ version=0, ) parsinlu_reading_comprehension_bigbench_lite = LightevalTaskConfig( - name="parsinlu_reading_comprehension", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:parsinlu_reading_comprehension", + suite=["lighteval"], prompt_function=prompt.bigbench_linefeed_before_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="parsinlu_reading_comprehension", @@ -15081,8 +15081,8 @@ version=0, ) penguins_in_a_table_bigbench = LightevalTaskConfig( - name="penguins_in_a_table", - suite=["bigbench", "bigbench_json"], + name="bigbench:penguins_in_a_table", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="penguins_in_a_table", @@ -15096,8 +15096,8 @@ version=0, ) periodic_elements_bigbench = LightevalTaskConfig( - name="periodic_elements", - suite=["bigbench", "bigbench_json"], + name="bigbench:periodic_elements", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="periodic_elements", @@ -15111,8 +15111,8 @@ version=0, ) persian_idioms_bigbench = LightevalTaskConfig( - name="persian_idioms", - suite=["bigbench", "bigbench_json"], + name="bigbench:persian_idioms", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="persian_idioms", @@ -15126,8 +15126,8 @@ version=0, ) phrase_relatedness_bigbench = LightevalTaskConfig( - name="phrase_relatedness", - suite=["bigbench", "bigbench_json"], + name="bigbench:phrase_relatedness", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="phrase_relatedness", @@ -15141,8 +15141,8 @@ version=0, ) physical_intuition_bigbench = LightevalTaskConfig( - name="physical_intuition", - suite=["bigbench", "bigbench_json"], + name="bigbench:physical_intuition", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="physical_intuition", @@ -15156,8 +15156,8 @@ version=0, ) physics_bigbench = LightevalTaskConfig( - name="physics", - suite=["bigbench", "bigbench_json"], + name="bigbench:physics", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="physics", @@ -15171,8 +15171,8 @@ version=0, ) physics_questions_bigbench = LightevalTaskConfig( - name="physics_questions", - suite=["bigbench", "bigbench_json"], + name="bigbench:physics_questions", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="physics_questions", @@ -15230,8 +15230,8 @@ version=0, ) play_dialog_same_or_different_bigbench_lite = LightevalTaskConfig( - name="play_dialog_same_or_different", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:play_dialog_same_or_different", + suite=["lighteval"], prompt_function=prompt.bigbench_linefeed_before_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="play_dialog_same_or_different", @@ -15245,8 +15245,8 @@ version=0, ) polish_sequence_labeling_bigbench = LightevalTaskConfig( - name="polish_sequence_labeling", - suite=["bigbench", "bigbench_json"], + name="bigbench:polish_sequence_labeling", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="polish_sequence_labeling", @@ -15260,8 +15260,8 @@ version=0, ) presuppositions_as_nli_bigbench = LightevalTaskConfig( - name="presuppositions_as_nli", - suite=["bigbench", "bigbench_json"], + name="bigbench:presuppositions_as_nli", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="presuppositions_as_nli", @@ -15388,8 +15388,8 @@ version=0, ) qa_wikidata_bigbench = LightevalTaskConfig( - name="qa_wikidata", - suite=["bigbench", "bigbench_json"], + name="bigbench:qa_wikidata", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="qa_wikidata", @@ -15457,8 +15457,8 @@ version=0, ) question_selection_bigbench = LightevalTaskConfig( - name="question_selection", - suite=["bigbench", "bigbench_json"], + name="bigbench:question_selection", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="question_selection", @@ -15795,8 +15795,8 @@ version=0, ) real_or_fake_text_bigbench = LightevalTaskConfig( - name="real_or_fake_text", - suite=["bigbench", "bigbench_json"], + name="bigbench:real_or_fake_text", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="real_or_fake_text", @@ -15825,8 +15825,8 @@ version=0, ) reasoning_about_colored_objects_bigbench = LightevalTaskConfig( - name="reasoning_about_colored_objects", - suite=["bigbench", "bigbench_json"], + name="bigbench:reasoning_about_colored_objects", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="reasoning_about_colored_objects", @@ -15840,8 +15840,8 @@ version=0, ) repeat_copy_logic_bigbench_lite = LightevalTaskConfig( - name="repeat_copy_logic", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:repeat_copy_logic", + suite=["lighteval"], prompt_function=prompt.bigbench_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="repeat_copy_logic", @@ -15855,8 +15855,8 @@ version=0, ) rephrase_bigbench = LightevalTaskConfig( - name="rephrase", - suite=["bigbench", "bigbench_json"], + name="bigbench:rephrase", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="rephrase", @@ -15875,8 +15875,8 @@ version=0, ) rhyming_bigbench = LightevalTaskConfig( - name="rhyming", - suite=["bigbench", "bigbench_json"], + name="bigbench:rhyming", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="rhyming", @@ -15890,8 +15890,8 @@ version=0, ) riddle_sense_bigbench = LightevalTaskConfig( - name="riddle_sense", - suite=["bigbench", "bigbench_json"], + name="bigbench:riddle_sense", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="riddle_sense", @@ -15905,8 +15905,8 @@ version=0, ) ruin_names_bigbench = LightevalTaskConfig( - name="ruin_names", - suite=["bigbench", "bigbench_json"], + name="bigbench:ruin_names", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="ruin_names", @@ -15920,8 +15920,8 @@ version=0, ) salient_translation_error_detection_bigbench = LightevalTaskConfig( - name="salient_translation_error_detection", - suite=["bigbench", "bigbench_json"], + name="bigbench:salient_translation_error_detection", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="salient_translation_error_detection", @@ -15935,8 +15935,8 @@ version=0, ) scientific_press_release_bigbench = LightevalTaskConfig( - name="scientific_press_release", - suite=["bigbench", "bigbench_json"], + name="bigbench:scientific_press_release", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="scientific_press_release", @@ -15968,8 +15968,8 @@ version=0, ) semantic_parsing_in_context_sparc_bigbench = LightevalTaskConfig( - name="semantic_parsing_in_context_sparc", - suite=["bigbench", "bigbench_json"], + name="bigbench:semantic_parsing_in_context_sparc", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="semantic_parsing_in_context_sparc", @@ -15983,8 +15983,8 @@ version=0, ) semantic_parsing_spider_bigbench = LightevalTaskConfig( - name="semantic_parsing_spider", - suite=["bigbench", "bigbench_json"], + name="bigbench:semantic_parsing_spider", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="semantic_parsing_spider", @@ -15998,8 +15998,8 @@ version=0, ) sentence_ambiguity_bigbench = LightevalTaskConfig( - name="sentence_ambiguity", - suite=["bigbench", "bigbench_json"], + name="bigbench:sentence_ambiguity", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="sentence_ambiguity", @@ -16013,8 +16013,8 @@ version=0, ) similarities_abstraction_bigbench = LightevalTaskConfig( - name="similarities_abstraction", - suite=["bigbench", "bigbench_json"], + name="bigbench:similarities_abstraction", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="similarities_abstraction", @@ -16028,8 +16028,8 @@ version=0, ) simp_turing_concept_bigbench = LightevalTaskConfig( - name="simp_turing_concept", - suite=["bigbench", "bigbench_json"], + name="bigbench:simp_turing_concept", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="simp_turing_concept", @@ -16058,8 +16058,8 @@ version=0, ) simple_arithmetic_json_bigbench = LightevalTaskConfig( - name="simple_arithmetic_json", - suite=["bigbench", "bigbench_json"], + name="bigbench:simple_arithmetic_json", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="simple_arithmetic_json", @@ -16073,8 +16073,8 @@ version=0, ) simple_arithmetic_json_multiple_choice_bigbench = LightevalTaskConfig( - name="simple_arithmetic_json_multiple_choice", - suite=["bigbench", "bigbench_json"], + name="bigbench:simple_arithmetic_json_multiple_choice", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="simple_arithmetic_json_multiple_choice", @@ -16088,8 +16088,8 @@ version=0, ) simple_arithmetic_json_subtasks_bigbench = LightevalTaskConfig( - name="simple_arithmetic_json_subtasks", - suite=["bigbench", "bigbench_json"], + name="bigbench:simple_arithmetic_json_subtasks", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="simple_arithmetic_json_subtasks", @@ -16103,8 +16103,8 @@ version=0, ) simple_arithmetic_multiple_targets_json_bigbench = LightevalTaskConfig( - name="simple_arithmetic_multiple_targets_json", - suite=["bigbench", "bigbench_json"], + name="bigbench:simple_arithmetic_multiple_targets_json", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="simple_arithmetic_multiple_targets_json", @@ -16118,8 +16118,8 @@ version=0, ) simple_ethical_questions_bigbench = LightevalTaskConfig( - name="simple_ethical_questions", - suite=["bigbench", "bigbench_json"], + name="bigbench:simple_ethical_questions", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="simple_ethical_questions", @@ -16133,8 +16133,8 @@ version=0, ) simple_text_editing_bigbench = LightevalTaskConfig( - name="simple_text_editing", - suite=["bigbench", "bigbench_json"], + name="bigbench:simple_text_editing", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="simple_text_editing", @@ -16174,8 +16174,8 @@ version=0, ) snarks_bigbench = LightevalTaskConfig( - name="snarks", - suite=["bigbench", "bigbench_json"], + name="bigbench:snarks", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="snarks", @@ -16189,8 +16189,8 @@ version=0, ) social_iqa_bigbench = LightevalTaskConfig( - name="social_iqa", - suite=["bigbench", "bigbench_json"], + name="bigbench:social_iqa", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="social_iqa", @@ -16204,8 +16204,8 @@ version=0, ) social_support_bigbench = LightevalTaskConfig( - name="social_support", - suite=["bigbench", "bigbench_json"], + name="bigbench:social_support", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="social_support", @@ -16219,8 +16219,8 @@ version=0, ) sports_understanding_bigbench = LightevalTaskConfig( - name="sports_understanding", - suite=["bigbench", "bigbench_json"], + name="bigbench:sports_understanding", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="sports_understanding", @@ -16293,8 +16293,8 @@ version=0, ) strange_stories_bigbench_lite = LightevalTaskConfig( - name="strange_stories", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:strange_stories", + suite=["lighteval"], prompt_function=prompt.bigbench_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="strange_stories", @@ -16308,8 +16308,8 @@ version=0, ) strategyqa_bigbench_lite = LightevalTaskConfig( - name="strategyqa", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:strategyqa", + suite=["lighteval"], prompt_function=prompt.bigbench_linefeed_before_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="strategyqa", @@ -16323,8 +16323,8 @@ version=0, ) sufficient_information_bigbench = LightevalTaskConfig( - name="sufficient_information", - suite=["bigbench", "bigbench_json"], + name="bigbench:sufficient_information", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="sufficient_information", @@ -16338,8 +16338,8 @@ version=0, ) suicide_risk_bigbench = LightevalTaskConfig( - name="suicide_risk", - suite=["bigbench", "bigbench_json"], + name="bigbench:suicide_risk", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="suicide_risk", @@ -16524,8 +16524,8 @@ version=0, ) swahili_english_proverbs_bigbench = LightevalTaskConfig( - name="swahili_english_proverbs", - suite=["bigbench", "bigbench_json"], + name="bigbench:swahili_english_proverbs", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="swahili_english_proverbs", @@ -16557,8 +16557,8 @@ version=0, ) swedish_to_german_proverbs_bigbench = LightevalTaskConfig( - name="swedish_to_german_proverbs", - suite=["bigbench", "bigbench_json"], + name="bigbench:swedish_to_german_proverbs", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="swedish_to_german_proverbs", @@ -16572,8 +16572,8 @@ version=0, ) symbol_interpretation_bigbench_lite = LightevalTaskConfig( - name="symbol_interpretation", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:symbol_interpretation", + suite=["lighteval"], prompt_function=prompt.bigbench_linefeed_before_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="symbol_interpretation", @@ -16695,8 +16695,8 @@ version=0, ) tellmewhy_bigbench = LightevalTaskConfig( - name="tellmewhy", - suite=["bigbench", "bigbench_json"], + name="bigbench:tellmewhy", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="tellmewhy", @@ -16710,8 +16710,8 @@ version=0, ) temporal_sequences_bigbench = LightevalTaskConfig( - name="temporal_sequences", - suite=["bigbench", "bigbench_json"], + name="bigbench:temporal_sequences", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="temporal_sequences", @@ -16725,8 +16725,8 @@ version=0, ) tense_bigbench = LightevalTaskConfig( - name="tense", - suite=["bigbench", "bigbench_json"], + name="bigbench:tense", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="tense", @@ -17025,8 +17025,8 @@ version=0, ) timedial_bigbench = LightevalTaskConfig( - name="timedial", - suite=["bigbench", "bigbench_json"], + name="bigbench:timedial", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="timedial", @@ -17058,8 +17058,8 @@ version=0, ) topical_chat_bigbench = LightevalTaskConfig( - name="topical_chat", - suite=["bigbench", "bigbench_json"], + name="bigbench:topical_chat", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="topical_chat", @@ -17073,8 +17073,8 @@ version=0, ) tracking_shuffled_objects_bigbench = LightevalTaskConfig( - name="tracking_shuffled_objects", - suite=["bigbench", "bigbench_json"], + name="bigbench:tracking_shuffled_objects", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="tracking_shuffled_objects", @@ -17192,8 +17192,8 @@ version=0, ) understanding_fables_bigbench = LightevalTaskConfig( - name="understanding_fables", - suite=["bigbench", "bigbench_json"], + name="bigbench:understanding_fables", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="understanding_fables", @@ -17207,8 +17207,8 @@ version=0, ) undo_permutation_bigbench = LightevalTaskConfig( - name="undo_permutation", - suite=["bigbench", "bigbench_json"], + name="bigbench:undo_permutation", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="undo_permutation", @@ -17222,8 +17222,8 @@ version=0, ) unit_conversion_bigbench = LightevalTaskConfig( - name="unit_conversion", - suite=["bigbench", "bigbench_json"], + name="bigbench:unit_conversion", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="unit_conversion", @@ -17237,8 +17237,8 @@ version=0, ) unit_interpretation_bigbench = LightevalTaskConfig( - name="unit_interpretation", - suite=["bigbench", "bigbench_json"], + name="bigbench:unit_interpretation", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="unit_interpretation", @@ -17252,8 +17252,8 @@ version=0, ) unnatural_in_context_learning_bigbench = LightevalTaskConfig( - name="unnatural_in_context_learning", - suite=["bigbench", "bigbench_json"], + name="bigbench:unnatural_in_context_learning", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="unnatural_in_context_learning", @@ -17342,8 +17342,8 @@ version=0, ) vitaminc_fact_verification_bigbench_lite = LightevalTaskConfig( - name="vitaminc_fact_verification", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:vitaminc_fact_verification", + suite=["lighteval"], prompt_function=prompt.bigbench_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="vitaminc_fact_verification", @@ -17372,8 +17372,8 @@ version=0, ) what_is_the_tao_bigbench = LightevalTaskConfig( - name="what_is_the_tao", - suite=["bigbench", "bigbench_json"], + name="bigbench:what_is_the_tao", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="what_is_the_tao", @@ -17387,8 +17387,8 @@ version=0, ) which_wiki_edit_bigbench = LightevalTaskConfig( - name="which_wiki_edit", - suite=["bigbench", "bigbench_json"], + name="bigbench:which_wiki_edit", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="which_wiki_edit", @@ -19683,8 +19683,8 @@ version=0, ) wino_x_german_bigbench = LightevalTaskConfig( - name="wino_x_german", - suite=["bigbench", "bigbench_json"], + name="bigbench:wino_x_german", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="wino_x_german", @@ -19713,8 +19713,8 @@ version=0, ) winowhy_bigbench_lite = LightevalTaskConfig( - name="winowhy", - suite=["bigbench_lite", "bigbench", "bigbench_json"], + name="bigbench_lite:winowhy", + suite=["lighteval"], prompt_function=prompt.bigbench_whitespace_after_query, hf_repo="tasksource/bigbench", hf_subset="winowhy", @@ -22248,8 +22248,8 @@ version=0, ) word_sorting_bigbench = LightevalTaskConfig( - name="word_sorting", - suite=["bigbench", "bigbench_json"], + name="bigbench:word_sorting", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="word_sorting", @@ -22263,8 +22263,8 @@ version=0, ) word_unscrambling_bigbench = LightevalTaskConfig( - name="word_unscrambling", - suite=["bigbench", "bigbench_json"], + name="bigbench:word_unscrambling", + suite=["lighteval"], prompt_function=prompt.bigbench, hf_repo="tasksource/bigbench", hf_subset="word_unscrambling",