summarization_cnndm:temperature=0.3,device=cuda,model=anthropic_stanford-online-all-v4-s3 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=cohere_command-medium-beta | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=cohere_command-xlarge-beta | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=cohere_large-20220720 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=cohere_medium-20220720 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=cohere_medium-20221108 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=cohere_small-20220720 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=cohere_xlarge-20220609 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=cohere_xlarge-20221108 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=microsoft_TNLGv2_530B | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=microsoft_TNLGv2_7B | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_ada | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_babbage | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_curie | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_curie,max_train_instances=0 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_davinci | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_davinci,max_train_instances=0 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_gpt-3.5-turbo-0301 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_text-ada-001 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_text-babbage-001 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_text-curie-001 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_text-curie-001,max_train_instances=0 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_text-davinci-002 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_text-davinci-002,max_train_instances=0 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=openai_text-davinci-003 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=together_bloom | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=together_glm,stop=hash | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=together_gpt-j-6b | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=together_gpt-neox-20b | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=together_opt-175b | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=together_opt-66b | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=together_redpajama-incite-base-3b-v1 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=together_t0pp,stop=hash | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=together_t5-11b,stop=hash | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=together_ul2,stop=hash,global_prefix=nlg | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=together_yalm | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=writer_palmyra-instruct-30 | generation |
summarization_cnndm:temperature=0.3,device=cuda,model=writer_palmyra-x | generation |
summarization_xsum:temperature=0.3,device=cpu,model=AlephAlpha_luminous-base | generation |
summarization_xsum:temperature=0.3,device=cpu,model=AlephAlpha_luminous-extended | generation |
summarization_xsum:temperature=0.3,device=cpu,model=AlephAlpha_luminous-supreme | generation |
summarization_xsum:temperature=0.3,device=cpu,model=ai21_j1-grande | generation |
summarization_xsum:temperature=0.3,device=cpu,model=ai21_j1-grande-v2-beta | generation |
summarization_xsum:temperature=0.3,device=cpu,model=ai21_j1-jumbo | generation |
summarization_xsum:temperature=0.3,device=cpu,model=ai21_j1-large | generation |
summarization_xsum:temperature=0.3,device=cpu,model=ai21_j2-grande | generation |
summarization_xsum:temperature=0.3,device=cpu,model=ai21_j2-jumbo | generation |
summarization_xsum:temperature=0.3,device=cpu,model=ai21_j2-large | generation |
summarization_xsum:temperature=0.3,device=cpu,model=anthropic_stanford-online-all-v4-s3 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=cohere_command-medium-beta | generation |
summarization_xsum:temperature=0.3,device=cpu,model=cohere_command-xlarge-beta | generation |
summarization_xsum:temperature=0.3,device=cpu,model=cohere_large-20220720 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=cohere_medium-20220720 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=cohere_medium-20221108 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=cohere_small-20220720 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=cohere_xlarge-20220609 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=cohere_xlarge-20221108 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=microsoft_TNLGv2_530B | generation |
summarization_xsum:temperature=0.3,device=cpu,model=microsoft_TNLGv2_7B | generation |
summarization_xsum:temperature=0.3,device=cpu,model=openai_ada | generation |
summarization_xsum:temperature=0.3,device=cpu,model=openai_babbage | generation |
summarization_xsum:temperature=0.3,device=cpu,model=openai_curie | generation |
summarization_xsum:temperature=0.3,device=cpu,model=openai_davinci | generation |
summarization_xsum:temperature=0.3,device=cpu,model=openai_gpt-3.5-turbo-0301 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=openai_text-ada-001 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=openai_text-babbage-001 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=openai_text-curie-001 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=openai_text-davinci-002 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=openai_text-davinci-003 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=together_bloom | generation |
summarization_xsum:temperature=0.3,device=cpu,model=together_glm,stop=hash | generation |
summarization_xsum:temperature=0.3,device=cpu,model=together_gpt-j-6b | generation |
summarization_xsum:temperature=0.3,device=cpu,model=together_gpt-neox-20b | generation |
summarization_xsum:temperature=0.3,device=cpu,model=together_opt-175b | generation |
summarization_xsum:temperature=0.3,device=cpu,model=together_opt-66b | generation |
summarization_xsum:temperature=0.3,device=cpu,model=together_redpajama-incite-base-3b-v1 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=together_t0pp,stop=hash | generation |
summarization_xsum:temperature=0.3,device=cpu,model=together_t5-11b,stop=hash | generation |
summarization_xsum:temperature=0.3,device=cpu,model=together_ul2,stop=hash,global_prefix=nlg | generation |
summarization_xsum:temperature=0.3,device=cpu,model=together_yalm | generation |
summarization_xsum:temperature=0.3,device=cpu,model=writer_palmyra-instruct-30 | generation |
summarization_xsum:temperature=0.3,device=cpu,model=writer_palmyra-x | generation |
summarization_xsum:temperature=0.3,device=cuda,model=AlephAlpha_luminous-base | generation |
summarization_xsum:temperature=0.3,device=cuda,model=AlephAlpha_luminous-extended | generation |
summarization_xsum:temperature=0.3,device=cuda,model=AlephAlpha_luminous-supreme | generation |
summarization_xsum:temperature=0.3,device=cuda,model=ai21_j1-grande | generation |
summarization_xsum:temperature=0.3,device=cuda,model=ai21_j1-grande-v2-beta | generation |
summarization_xsum:temperature=0.3,device=cuda,model=ai21_j1-jumbo | generation |
summarization_xsum:temperature=0.3,device=cuda,model=ai21_j1-large | generation |
summarization_xsum:temperature=0.3,device=cuda,model=ai21_j2-grande | generation |
summarization_xsum:temperature=0.3,device=cuda,model=ai21_j2-jumbo | generation |
summarization_xsum:temperature=0.3,device=cuda,model=ai21_j2-large | generation |
summarization_xsum:temperature=0.3,device=cuda,model=anthropic_stanford-online-all-v4-s3 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=cohere_command-medium-beta | generation |
summarization_xsum:temperature=0.3,device=cuda,model=cohere_command-xlarge-beta | generation |
summarization_xsum:temperature=0.3,device=cuda,model=cohere_large-20220720 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=cohere_medium-20220720 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=cohere_medium-20221108 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=cohere_small-20220720 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=cohere_xlarge-20220609 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=cohere_xlarge-20221108 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=microsoft_TNLGv2_530B | generation |
summarization_xsum:temperature=0.3,device=cuda,model=microsoft_TNLGv2_7B | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_ada | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_babbage | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_curie | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_curie,max_train_instances=0 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_davinci | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_davinci,max_train_instances=0 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_gpt-3.5-turbo-0301 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_text-ada-001 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_text-babbage-001 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_text-curie-001 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_text-curie-001,max_train_instances=0 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_text-davinci-002 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_text-davinci-002,max_train_instances=0 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=openai_text-davinci-003 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=together_bloom | generation |
summarization_xsum:temperature=0.3,device=cuda,model=together_glm,stop=hash | generation |
summarization_xsum:temperature=0.3,device=cuda,model=together_gpt-j-6b | generation |
summarization_xsum:temperature=0.3,device=cuda,model=together_gpt-neox-20b | generation |
summarization_xsum:temperature=0.3,device=cuda,model=together_opt-175b | generation |
summarization_xsum:temperature=0.3,device=cuda,model=together_opt-66b | generation |
summarization_xsum:temperature=0.3,device=cuda,model=together_redpajama-incite-base-3b-v1 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=together_t0pp,stop=hash | generation |
summarization_xsum:temperature=0.3,device=cuda,model=together_t5-11b,stop=hash | generation |
summarization_xsum:temperature=0.3,device=cuda,model=together_ul2,stop=hash,global_prefix=nlg | generation |
summarization_xsum:temperature=0.3,device=cuda,model=together_yalm | generation |
summarization_xsum:temperature=0.3,device=cuda,model=writer_palmyra-instruct-30 | generation |
summarization_xsum:temperature=0.3,device=cuda,model=writer_palmyra-x | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1024,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=1536,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=256,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=ai21_j1-grande-v2-beta,tokenizer=ai21_j1,num_prompt_tokens=512,num_output_tokens=8 | generation |