task_id stringlengths 8 69 | name stringlengths 3 64 | suite stringclasses 6 values | hf_repo stringclasses 125 values | hf_subset stringlengths 0 55 | file_path stringclasses 6 values | line_number int64 36 22.8k | variable_name stringclasses 1 value | is_subtask bool 2 classes | main_task stringclasses 122 values | subtask_count int64 1 1 | suites listlengths 1 1 | subtasks listlengths 1 1 | is_standalone bool 2 classes |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
lighteval:mmmu_pro:standard-4 | mmmu_pro:standard-4 | lighteval | MMMU/MMMU_pro | standard (4 options) | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 36 | direct_call | true | mmmu_pro | 1 | [
"lighteval"
] | [
"lighteval:mmmu_pro:standard-4"
] | false |
lighteval:mmmu_pro:standard-10 | mmmu_pro:standard-10 | lighteval | MMMU/MMMU_pro | standard (10 options) | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 51 | direct_call | true | mmmu_pro | 1 | [
"lighteval"
] | [
"lighteval:mmmu_pro:standard-10"
] | false |
lighteval:mmmu_pro:vision | mmmu_pro:vision | lighteval | MMMU/MMMU_pro | vision | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 66 | direct_call | true | mmmu_pro | 1 | [
"lighteval"
] | [
"lighteval:mmmu_pro:vision"
] | false |
lighteval:bigbench:abstract_narrative_understanding | bigbench:abstract_narrative_understanding | lighteval | tasksource/bigbench | abstract_narrative_understanding | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 81 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:abstract_narrative_understanding"
] | false |
lighteval:agieval:aqua-rat | agieval:aqua-rat | lighteval | dmayhem93/agieval-aqua-rat | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 96 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:aqua-rat"
] | false |
lighteval:agieval:gaokao-biology | agieval:gaokao-biology | lighteval | dmayhem93/agieval-gaokao-biology | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 114 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:gaokao-biology"
] | false |
lighteval:agieval:gaokao-chemistry | agieval:gaokao-chemistry | lighteval | dmayhem93/agieval-gaokao-chemistry | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 132 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:gaokao-chemistry"
] | false |
lighteval:agieval:gaokao-chinese | agieval:gaokao-chinese | lighteval | dmayhem93/agieval-gaokao-chinese | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 150 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:gaokao-chinese"
] | false |
lighteval:agieval:gaokao-english | agieval:gaokao-english | lighteval | dmayhem93/agieval-gaokao-english | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 168 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:gaokao-english"
] | false |
lighteval:agieval:gaokao-geography | agieval:gaokao-geography | lighteval | dmayhem93/agieval-gaokao-geography | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 186 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:gaokao-geography"
] | false |
lighteval:agieval:gaokao-history | agieval:gaokao-history | lighteval | dmayhem93/agieval-gaokao-history | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 204 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:gaokao-history"
] | false |
lighteval:agieval:gaokao-mathqa | agieval:gaokao-mathqa | lighteval | dmayhem93/agieval-gaokao-mathqa | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 222 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:gaokao-mathqa"
] | false |
lighteval:agieval:gaokao-physics | agieval:gaokao-physics | lighteval | dmayhem93/agieval-gaokao-physics | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 240 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:gaokao-physics"
] | false |
lighteval:agieval:logiqa-en | agieval:logiqa-en | lighteval | dmayhem93/agieval-logiqa-en | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 258 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:logiqa-en"
] | false |
lighteval:agieval:logiqa-zh | agieval:logiqa-zh | lighteval | dmayhem93/agieval-logiqa-zh | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 276 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:logiqa-zh"
] | false |
lighteval:agieval:lsat-ar | agieval:lsat-ar | lighteval | dmayhem93/agieval-lsat-ar | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 294 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:lsat-ar"
] | false |
lighteval:agieval:lsat-lr | agieval:lsat-lr | lighteval | dmayhem93/agieval-lsat-lr | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 312 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:lsat-lr"
] | false |
lighteval:agieval:lsat-rc | agieval:lsat-rc | lighteval | dmayhem93/agieval-lsat-rc | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 330 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:lsat-rc"
] | false |
lighteval:agieval:sat-en | agieval:sat-en | lighteval | dmayhem93/agieval-sat-en | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 348 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:sat-en"
] | false |
lighteval:agieval:sat-en-without-passage | agieval:sat-en-without-passage | lighteval | dmayhem93/agieval-sat-en-without-passage | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 366 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:sat-en-without-passage"
] | false |
lighteval:agieval:sat-math | agieval:sat-math | lighteval | dmayhem93/agieval-sat-math | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 384 | direct_call | true | agieval | 1 | [
"lighteval"
] | [
"lighteval:agieval:sat-math"
] | false |
lighteval:aime24 | aime24 | lighteval | HuggingFaceH4/aime_2024 | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 402 | direct_call | false | aime24 | 1 | [
"lighteval"
] | [
"lighteval:aime24"
] | true |
lighteval:aime24_avg | aime24_avg | lighteval | HuggingFaceH4/aime_2024 | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 416 | direct_call | false | aime24_avg | 1 | [
"lighteval"
] | [
"lighteval:aime24_avg"
] | true |
lighteval:aime24_gpassk | aime24_gpassk | lighteval | HuggingFaceH4/aime_2024 | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 430 | direct_call | false | aime24_gpassk | 1 | [
"lighteval"
] | [
"lighteval:aime24_gpassk"
] | true |
lighteval:aime25 | aime25 | lighteval | yentinglin/aime_2025 | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 444 | direct_call | false | aime25 | 1 | [
"lighteval"
] | [
"lighteval:aime25"
] | true |
lighteval:aime25_gpassk | aime25_gpassk | lighteval | yentinglin/aime_2025 | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 458 | direct_call | false | aime25_gpassk | 1 | [
"lighteval"
] | [
"lighteval:aime25_gpassk"
] | true |
lighteval:bigbench:anachronisms | bigbench:anachronisms | lighteval | tasksource/bigbench | anachronisms | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 472 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:anachronisms"
] | false |
lighteval:bigbench:analogical_similarity | bigbench:analogical_similarity | lighteval | tasksource/bigbench | analogical_similarity | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 487 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:analogical_similarity"
] | false |
lighteval:bigbench:analytic_entailment | bigbench:analytic_entailment | lighteval | tasksource/bigbench | analytic_entailment | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 502 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:analytic_entailment"
] | false |
lighteval:anli:r1 | anli:r1 | lighteval | anli | plain_text | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 517 | direct_call | true | anli | 1 | [
"lighteval"
] | [
"lighteval:anli:r1"
] | false |
lighteval:anli:r2 | anli:r2 | lighteval | anli | plain_text | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 532 | direct_call | true | anli | 1 | [
"lighteval"
] | [
"lighteval:anli:r2"
] | false |
lighteval:anli:r3 | anli:r3 | lighteval | anli | plain_text | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 547 | direct_call | true | anli | 1 | [
"lighteval"
] | [
"lighteval:anli:r3"
] | false |
lighteval:arc_agi_2 | arc_agi_2 | lighteval | arc-agi-community/arc-agi-2 | default | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 562 | direct_call | false | arc_agi_2 | 1 | [
"lighteval"
] | [
"lighteval:arc_agi_2"
] | true |
original:arc:c:letters | arc:c:letters | original | ai2_arc | ARC-Challenge | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 577 | direct_call | true | arc | 1 | [
"original"
] | [
"original:arc:c:letters"
] | false |
original:arc:c:options | arc:c:options | original | ai2_arc | ARC-Challenge | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 592 | direct_call | true | arc | 1 | [
"original"
] | [
"original:arc:c:options"
] | false |
original:arc:c:simple | arc:c:simple | original | ai2_arc | ARC-Challenge | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 610 | direct_call | true | arc | 1 | [
"original"
] | [
"original:arc:c:simple"
] | false |
leaderboard:arc:challenge | arc:challenge | leaderboard | ai2_arc | ARC-Challenge | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 628 | direct_call | true | arc | 1 | [
"leaderboard"
] | [
"leaderboard:arc:challenge"
] | false |
lighteval:arc:easy | arc:easy | lighteval | ai2_arc | ARC-Easy | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 646 | direct_call | true | arc | 1 | [
"lighteval"
] | [
"lighteval:arc:easy"
] | false |
lighteval:arithmetic:1dc | arithmetic:1dc | lighteval | EleutherAI/arithmetic | arithmetic_1dc | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 664 | direct_call | true | arithmetic | 1 | [
"lighteval"
] | [
"lighteval:arithmetic:1dc"
] | false |
lighteval:arithmetic:2da | arithmetic:2da | lighteval | EleutherAI/arithmetic | arithmetic_2da | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 679 | direct_call | true | arithmetic | 1 | [
"lighteval"
] | [
"lighteval:arithmetic:2da"
] | false |
lighteval:arithmetic:2dm | arithmetic:2dm | lighteval | EleutherAI/arithmetic | arithmetic_2dm | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 694 | direct_call | true | arithmetic | 1 | [
"lighteval"
] | [
"lighteval:arithmetic:2dm"
] | false |
lighteval:arithmetic:2ds | arithmetic:2ds | lighteval | EleutherAI/arithmetic | arithmetic_2ds | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 709 | direct_call | true | arithmetic | 1 | [
"lighteval"
] | [
"lighteval:arithmetic:2ds"
] | false |
lighteval:arithmetic:3da | arithmetic:3da | lighteval | EleutherAI/arithmetic | arithmetic_3da | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 724 | direct_call | true | arithmetic | 1 | [
"lighteval"
] | [
"lighteval:arithmetic:3da"
] | false |
lighteval:arithmetic:3ds | arithmetic:3ds | lighteval | EleutherAI/arithmetic | arithmetic_3ds | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 739 | direct_call | true | arithmetic | 1 | [
"lighteval"
] | [
"lighteval:arithmetic:3ds"
] | false |
lighteval:arithmetic:4da | arithmetic:4da | lighteval | EleutherAI/arithmetic | arithmetic_4da | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 754 | direct_call | true | arithmetic | 1 | [
"lighteval"
] | [
"lighteval:arithmetic:4da"
] | false |
lighteval:arithmetic:4ds | arithmetic:4ds | lighteval | EleutherAI/arithmetic | arithmetic_4ds | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 769 | direct_call | true | arithmetic | 1 | [
"lighteval"
] | [
"lighteval:arithmetic:4ds"
] | false |
lighteval:arithmetic:5da | arithmetic:5da | lighteval | EleutherAI/arithmetic | arithmetic_5da | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 784 | direct_call | true | arithmetic | 1 | [
"lighteval"
] | [
"lighteval:arithmetic:5da"
] | false |
lighteval:arithmetic:5ds | arithmetic:5ds | lighteval | EleutherAI/arithmetic | arithmetic_5ds | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 799 | direct_call | true | arithmetic | 1 | [
"lighteval"
] | [
"lighteval:arithmetic:5ds"
] | false |
lighteval:bigbench:arithmetic_bb | bigbench:arithmetic_bb | lighteval | tasksource/bigbench | arithmetic | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 814 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:arithmetic_bb"
] | false |
lighteval:bigbench:ascii_word_recognition | bigbench:ascii_word_recognition | lighteval | tasksource/bigbench | ascii_word_recognition | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 829 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:ascii_word_recognition"
] | false |
lighteval:asdiv | asdiv | lighteval | EleutherAI/asdiv | asdiv | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 844 | direct_call | false | asdiv | 1 | [
"lighteval"
] | [
"lighteval:asdiv"
] | true |
lighteval:bigbench:authorship_verification | bigbench:authorship_verification | lighteval | tasksource/bigbench | authorship_verification | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 859 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:authorship_verification"
] | false |
lighteval:bigbench:auto_categorization | bigbench:auto_categorization | lighteval | tasksource/bigbench | auto_categorization | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 874 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:auto_categorization"
] | false |
lighteval:bigbench_lite:auto_debugging | bigbench_lite:auto_debugging | lighteval | tasksource/bigbench | auto_debugging | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 889 | direct_call | true | bigbench_lite | 1 | [
"lighteval"
] | [
"lighteval:bigbench_lite:auto_debugging"
] | false |
helm:babi_qa | babi_qa | helm | facebook/babi_qa | en-valid-qa1 | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 904 | direct_call | false | babi_qa | 1 | [
"helm"
] | [
"helm:babi_qa"
] | true |
lighteval:bigbench:causal_judgment | bigbench:causal_judgment | lighteval | lighteval/bbh | causal_judgement | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 930 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:causal_judgment"
] | false |
lighteval:bigbench:date_understanding | bigbench:date_understanding | lighteval | lighteval/bbh | date_understanding | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 945 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:date_understanding"
] | false |
lighteval:bigbench:disambiguation_qa | bigbench:disambiguation_qa | lighteval | lighteval/bbh | disambiguation_qa | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 960 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:disambiguation_qa"
] | false |
lighteval:bigbench:geometric_shapes | bigbench:geometric_shapes | lighteval | lighteval/bbh | geometric_shapes | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 975 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:geometric_shapes"
] | false |
lighteval:bigbench:logical_deduction_five_objects | bigbench:logical_deduction_five_objects | lighteval | lighteval/bbh | logical_deduction_five_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 990 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:logical_deduction_five_objects"
] | false |
lighteval:bigbench:logical_deduction_seven_objects | bigbench:logical_deduction_seven_objects | lighteval | lighteval/bbh | logical_deduction_seven_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,005 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:logical_deduction_seven_objects"
] | false |
lighteval:bigbench:logical_deduction_three_objects | bigbench:logical_deduction_three_objects | lighteval | lighteval/bbh | logical_deduction_three_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,020 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:logical_deduction_three_objects"
] | false |
lighteval:bigbench:movie_recommendation | bigbench:movie_recommendation | lighteval | lighteval/bbh | movie_recommendation | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,035 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:movie_recommendation"
] | false |
lighteval:bigbench:navigate | bigbench:navigate | lighteval | lighteval/bbh | navigate | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,050 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:navigate"
] | false |
lighteval:bigbench:reasoning_about_colored_objects | bigbench:reasoning_about_colored_objects | lighteval | lighteval/bbh | reasoning_about_colored_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,065 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:reasoning_about_colored_objects"
] | false |
lighteval:bigbench:ruin_names | bigbench:ruin_names | lighteval | lighteval/bbh | ruin_names | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,080 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:ruin_names"
] | false |
lighteval:bigbench:salient_translation_error_detection | bigbench:salient_translation_error_detection | lighteval | lighteval/bbh | salient_translation_error_detection | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,095 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:salient_translation_error_detection"
] | false |
lighteval:bigbench:snarks | bigbench:snarks | lighteval | lighteval/bbh | snarks | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,110 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:snarks"
] | false |
lighteval:bigbench:sports_understanding | bigbench:sports_understanding | lighteval | lighteval/bbh | sports_understanding | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,125 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:sports_understanding"
] | false |
lighteval:bigbench:temporal_sequences | bigbench:temporal_sequences | lighteval | lighteval/bbh | temporal_sequences | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,140 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:temporal_sequences"
] | false |
lighteval:bigbench:tracking_shuffled_objects_five_objects | bigbench:tracking_shuffled_objects_five_objects | lighteval | lighteval/bbh | tracking_shuffled_objects_five_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,155 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:tracking_shuffled_objects_five_objects"
] | false |
lighteval:bigbench:tracking_shuffled_objects_seven_objects | bigbench:tracking_shuffled_objects_seven_objects | lighteval | lighteval/bbh | tracking_shuffled_objects_seven_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,170 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:tracking_shuffled_objects_seven_objects"
] | false |
lighteval:bigbench:tracking_shuffled_objects_three_objects | bigbench:tracking_shuffled_objects_three_objects | lighteval | lighteval/bbh | tracking_shuffled_objects_three_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,185 | direct_call | true | bigbench | 1 | [
"lighteval"
] | [
"lighteval:bigbench:tracking_shuffled_objects_three_objects"
] | false |
harness:bigbench:causal_judgment | bigbench:causal_judgment | harness | lighteval/bbh | causal_judgement | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,200 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:causal_judgment"
] | false |
harness:bigbench:date_understanding | bigbench:date_understanding | harness | lighteval/bbh | date_understanding | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,219 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:date_understanding"
] | false |
harness:bigbench:disambiguation_qa | bigbench:disambiguation_qa | harness | lighteval/bbh | disambiguation_qa | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,238 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:disambiguation_qa"
] | false |
harness:bigbench:geometric_shapes | bigbench:geometric_shapes | harness | lighteval/bbh | geometric_shapes | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,257 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:geometric_shapes"
] | false |
harness:bigbench:logical_deduction_five_objects | bigbench:logical_deduction_five_objects | harness | lighteval/bbh | logical_deduction_five_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,276 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:logical_deduction_five_objects"
] | false |
harness:bigbench:logical_deduction_seven_objects | bigbench:logical_deduction_seven_objects | harness | lighteval/bbh | logical_deduction_seven_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,295 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:logical_deduction_seven_objects"
] | false |
harness:bigbench:logical_deduction_three_objects | bigbench:logical_deduction_three_objects | harness | lighteval/bbh | logical_deduction_three_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,314 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:logical_deduction_three_objects"
] | false |
harness:bigbench:movie_recommendation | bigbench:movie_recommendation | harness | lighteval/bbh | movie_recommendation | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,333 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:movie_recommendation"
] | false |
harness:bigbench:navigate | bigbench:navigate | harness | lighteval/bbh | navigate | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,352 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:navigate"
] | false |
harness:bigbench:reasoning_about_colored_objects | bigbench:reasoning_about_colored_objects | harness | lighteval/bbh | reasoning_about_colored_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,371 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:reasoning_about_colored_objects"
] | false |
harness:bigbench:ruin_names | bigbench:ruin_names | harness | lighteval/bbh | ruin_names | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,390 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:ruin_names"
] | false |
harness:bigbench:salient_translation_error_detection | bigbench:salient_translation_error_detection | harness | lighteval/bbh | salient_translation_error_detection | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,409 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:salient_translation_error_detection"
] | false |
harness:bigbench:snarks | bigbench:snarks | harness | lighteval/bbh | snarks | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,428 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:snarks"
] | false |
harness:bigbench:sports_understanding | bigbench:sports_understanding | harness | lighteval/bbh | sports_understanding | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,447 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:sports_understanding"
] | false |
harness:bigbench:temporal_sequences | bigbench:temporal_sequences | harness | lighteval/bbh | temporal_sequences | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,466 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:temporal_sequences"
] | false |
harness:bigbench:tracking_shuffled_objects_five_objects | bigbench:tracking_shuffled_objects_five_objects | harness | lighteval/bbh | tracking_shuffled_objects_five_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,485 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:tracking_shuffled_objects_five_objects"
] | false |
harness:bigbench:tracking_shuffled_objects_seven_objects | bigbench:tracking_shuffled_objects_seven_objects | harness | lighteval/bbh | tracking_shuffled_objects_seven_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,504 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:tracking_shuffled_objects_seven_objects"
] | false |
harness:bigbench:tracking_shuffled_objects_three_objects | bigbench:tracking_shuffled_objects_three_objects | harness | lighteval/bbh | tracking_shuffled_objects_three_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,523 | direct_call | true | bigbench | 1 | [
"harness"
] | [
"harness:bigbench:tracking_shuffled_objects_three_objects"
] | false |
harness:bbh:boolean_expressions | bbh:boolean_expressions | harness | lukaemon/bbh | boolean_expressions | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,542 | direct_call | true | bbh | 1 | [
"harness"
] | [
"harness:bbh:boolean_expressions"
] | false |
harness:bbh:causal_judgment | bbh:causal_judgment | harness | lukaemon/bbh | causal_judgement | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,569 | direct_call | true | bbh | 1 | [
"harness"
] | [
"harness:bbh:causal_judgment"
] | false |
harness:bbh:date_understanding | bbh:date_understanding | harness | lukaemon/bbh | date_understanding | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,596 | direct_call | true | bbh | 1 | [
"harness"
] | [
"harness:bbh:date_understanding"
] | false |
harness:bbh:disambiguation_qa | bbh:disambiguation_qa | harness | lukaemon/bbh | disambiguation_qa | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,623 | direct_call | true | bbh | 1 | [
"harness"
] | [
"harness:bbh:disambiguation_qa"
] | false |
harness:bbh:dyck_languages | bbh:dyck_languages | harness | lukaemon/bbh | dyck_languages | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,650 | direct_call | true | bbh | 1 | [
"harness"
] | [
"harness:bbh:dyck_languages"
] | false |
harness:bbh:formal_fallacies | bbh:formal_fallacies | harness | lukaemon/bbh | formal_fallacies | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,677 | direct_call | true | bbh | 1 | [
"harness"
] | [
"harness:bbh:formal_fallacies"
] | false |
harness:bbh:geometric_shapes | bbh:geometric_shapes | harness | lukaemon/bbh | geometric_shapes | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,704 | direct_call | true | bbh | 1 | [
"harness"
] | [
"harness:bbh:geometric_shapes"
] | false |
harness:bbh:hyperbaton | bbh:hyperbaton | harness | lukaemon/bbh | hyperbaton | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,731 | direct_call | true | bbh | 1 | [
"harness"
] | [
"harness:bbh:hyperbaton"
] | false |
harness:bbh:logical_deduction_five_objects | bbh:logical_deduction_five_objects | harness | lukaemon/bbh | logical_deduction_five_objects | /Users/nathan/Repos/lighteval/src/lighteval/tasks/default_tasks.py | 1,758 | direct_call | true | bbh | 1 | [
"harness"
] | [
"harness:bbh:logical_deduction_five_objects"
] | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.