narrative_qa:model=openai_babbage,data_augmentation=canonical | generation |
narrative_qa:model=openai_curie,data_augmentation=canonical | generation |
narrative_qa:model=openai_davinci,data_augmentation=canonical | generation |
narrative_qa:model=openai_gpt-3.5-turbo-0301,data_augmentation=canonical | generation |
narrative_qa:model=openai_text-ada-001,data_augmentation=canonical | generation |
narrative_qa:model=openai_text-babbage-001,data_augmentation=canonical | generation |
narrative_qa:model=openai_text-curie-001,data_augmentation=canonical | generation |
narrative_qa:model=openai_text-davinci-002,data_augmentation=canonical | generation |
narrative_qa:model=openai_text-davinci-003,data_augmentation=canonical | generation |
narrative_qa:model=together_bloom,data_augmentation=canonical | generation |
narrative_qa:model=together_glm,data_augmentation=canonical,stop=hash | generation |
narrative_qa:model=together_gpt-j-6b,data_augmentation=canonical | generation |
narrative_qa:model=together_gpt-neox-20b,data_augmentation=canonical | generation |
narrative_qa:model=together_opt-175b,data_augmentation=canonical | generation |
narrative_qa:model=together_opt-66b,data_augmentation=canonical | generation |
narrative_qa:model=together_redpajama-incite-base-3b-v1,data_augmentation=canonical | generation |
narrative_qa:model=together_t0pp,data_augmentation=canonical,stop=hash | generation |
narrative_qa:model=together_t5-11b,data_augmentation=canonical,stop=hash | generation |
narrative_qa:model=together_ul2,data_augmentation=canonical,stop=hash,global_prefix=nlg | generation |
narrative_qa:model=together_yalm,data_augmentation=canonical | generation |
narrative_qa:model=writer_palmyra-instruct-30,data_augmentation=canonical | generation |
narrative_qa:model=writer_palmyra-x,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=AlephAlpha_luminous-base,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=AlephAlpha_luminous-extended,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=AlephAlpha_luminous-supreme,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=ai21_j1-grande,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=ai21_j1-grande-v2-beta,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=ai21_j1-jumbo,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=ai21_j1-large,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=ai21_j2-grande,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=ai21_j2-jumbo,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=ai21_j2-large,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=anthropic_stanford-online-all-v4-s3,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=cohere_command-medium-beta,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=cohere_command-xlarge-beta,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=cohere_large-20220720,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=cohere_medium-20220720,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=cohere_medium-20221108,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=cohere_small-20220720,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=cohere_xlarge-20220609,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=cohere_xlarge-20221108,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=microsoft_TNLGv2_530B,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=microsoft_TNLGv2_7B,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=openai_ada,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=openai_babbage,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=openai_curie,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=openai_davinci,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=openai_gpt-3.5-turbo-0301,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=openai_text-ada-001,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=openai_text-babbage-001,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=openai_text-curie-001,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=openai_text-davinci-002,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=openai_text-davinci-003,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=together_bloom,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=together_glm,data_augmentation=canonical,stop=hash | generation |
natural_qa:mode=closedbook,model=together_gpt-j-6b,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=together_gpt-neox-20b,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=together_opt-175b,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=together_opt-66b,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=together_redpajama-incite-base-3b-v1,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=together_t0pp,data_augmentation=canonical,stop=hash | generation |
natural_qa:mode=closedbook,model=together_t5-11b,data_augmentation=canonical,stop=hash | generation |
natural_qa:mode=closedbook,model=together_ul2,data_augmentation=canonical,stop=hash,global_prefix=nlg | generation |
natural_qa:mode=closedbook,model=together_yalm,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=writer_palmyra-instruct-30,data_augmentation=canonical | generation |
natural_qa:mode=closedbook,model=writer_palmyra-x,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=AlephAlpha_luminous-base,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=AlephAlpha_luminous-extended,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=AlephAlpha_luminous-supreme,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=ai21_j1-grande,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=ai21_j1-grande-v2-beta,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=ai21_j1-jumbo,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=ai21_j1-large,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=ai21_j2-grande,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=ai21_j2-jumbo,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=ai21_j2-large,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,instructions=expert,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,instructions=none,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,max_train_instances=0,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,max_train_instances=1,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,max_train_instances=16,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,max_train_instances=2,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,max_train_instances=4,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,max_train_instances=8,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,prompt=i_o,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,prompt=input_output,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=anthropic_stanford-online-all-v4-s3,prompt=input_output_html,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=cohere_command-medium-beta,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=cohere_command-xlarge-beta,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=cohere_large-20220720,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=cohere_medium-20220720,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=cohere_medium-20221108,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=cohere_small-20220720,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=cohere_xlarge-20220609,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=cohere_xlarge-20221108,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=microsoft_TNLGv2_530B,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=microsoft_TNLGv2_7B,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=openai_ada,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=openai_babbage,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=openai_curie,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=openai_davinci,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=openai_gpt-3.5-turbo-0301,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=openai_text-ada-001,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=openai_text-babbage-001,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=openai_text-curie-001,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=openai_text-davinci-002,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=openai_text-davinci-003,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=together_bloom,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=together_bloom,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_bloom,instructions=expert,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_bloom,instructions=none,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_bloom,max_train_instances=0,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_bloom,max_train_instances=1,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_bloom,max_train_instances=16,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_bloom,max_train_instances=2,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_bloom,max_train_instances=4,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_bloom,max_train_instances=8,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_bloom,prompt=i_o,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_bloom,prompt=input_output,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_bloom,prompt=input_output_html,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_glm,data_augmentation=canonical,stop=hash | generation |
natural_qa:mode=openbook_longans,model=together_glm,instructions=expert,stop=hash,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_glm,instructions=none,stop=hash,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_glm,max_train_instances=0,stop=hash,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_glm,max_train_instances=1,stop=hash,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_glm,max_train_instances=16,stop=hash,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_glm,max_train_instances=2,stop=hash,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_glm,max_train_instances=4,stop=hash,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_glm,max_train_instances=8,stop=hash,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_glm,prompt=i_o,stop=hash,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_glm,prompt=input_output,stop=hash,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_glm,prompt=input_output_html,stop=hash,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_glm,stop=hash,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,instructions=expert,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,instructions=none,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,max_train_instances=0,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,max_train_instances=1,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,max_train_instances=16,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,max_train_instances=2,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,max_train_instances=4,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,max_train_instances=8,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,prompt=i_o,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,prompt=input_output,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_gpt-j-6b,prompt=input_output_html,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,instructions=expert,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,instructions=none,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,max_train_instances=0,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,max_train_instances=1,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,max_train_instances=16,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,max_train_instances=2,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,max_train_instances=4,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,max_train_instances=8,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,prompt=i_o,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,prompt=input_output,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_gpt-neox-20b,prompt=input_output_html,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,instructions=expert,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,instructions=none,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,max_train_instances=0,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,max_train_instances=1,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,max_train_instances=16,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,max_train_instances=2,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,max_train_instances=4,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,max_train_instances=8,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,prompt=i_o,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,prompt=input_output,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_opt-175b,prompt=input_output_html,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,instructions=expert,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,instructions=none,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,max_train_instances=0,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,max_train_instances=1,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,max_train_instances=16,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,max_train_instances=2,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,max_train_instances=4,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,max_train_instances=8,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,prompt=i_o,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,prompt=input_output,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_opt-66b,prompt=input_output_html,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_redpajama-incite-base-3b-v1,data_augmentation=canonical | generation |
natural_qa:mode=openbook_longans,model=together_t0pp,data_augmentation=canonical,stop=hash | generation |
natural_qa:mode=openbook_longans,model=together_t0pp,instructions=expert,stop=hash,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_t0pp,instructions=none,stop=hash,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_t0pp,max_train_instances=0,stop=hash,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_t0pp,max_train_instances=1,stop=hash,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_t0pp,max_train_instances=16,stop=hash,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_t0pp,max_train_instances=2,stop=hash,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_t0pp,max_train_instances=4,stop=hash,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_t0pp,max_train_instances=8,stop=hash,groups=ablation_in_context | generation |
natural_qa:mode=openbook_longans,model=together_t0pp,prompt=i_o,stop=hash,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_t0pp,prompt=input_output,stop=hash,groups=ablation_prompts | generation |
natural_qa:mode=openbook_longans,model=together_t0pp,prompt=input_output_html,stop=hash,groups=ablation_prompts | generation |