synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=1024,num_output_tokens=32,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=1024,num_output_tokens=4,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=1024,num_output_tokens=64,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=1024,num_output_tokens=8,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=1536,num_output_tokens=1,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=1536,num_output_tokens=16,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=1536,num_output_tokens=2,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=1536,num_output_tokens=32,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=1536,num_output_tokens=4,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=1536,num_output_tokens=64,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=1536,num_output_tokens=8,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=256,num_output_tokens=1,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=256,num_output_tokens=16,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=256,num_output_tokens=2,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=256,num_output_tokens=32,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=256,num_output_tokens=4,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=256,num_output_tokens=64,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=256,num_output_tokens=8,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=512,num_output_tokens=1,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=512,num_output_tokens=16,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=512,num_output_tokens=2,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=512,num_output_tokens=32,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=512,num_output_tokens=4,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=512,num_output_tokens=64,stop=hash | generation |
synthetic_efficiency:random=None,model=together_t5-11b,tokenizer=google_t5,num_prompt_tokens=512,num_output_tokens=8,stop=hash | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1,num_output_tokens=1,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1,num_output_tokens=16,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1,num_output_tokens=2,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1,num_output_tokens=32,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1,num_output_tokens=4,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1,num_output_tokens=64,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1,num_output_tokens=8,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1024,num_output_tokens=1,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1024,num_output_tokens=16,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1024,num_output_tokens=2,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1024,num_output_tokens=32,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1024,num_output_tokens=4,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1024,num_output_tokens=64,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1024,num_output_tokens=8,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1536,num_output_tokens=1,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1536,num_output_tokens=16,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1536,num_output_tokens=2,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1536,num_output_tokens=32,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1536,num_output_tokens=4,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1536,num_output_tokens=64,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=1536,num_output_tokens=8,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=256,num_output_tokens=1,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=256,num_output_tokens=16,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=256,num_output_tokens=2,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=256,num_output_tokens=32,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=256,num_output_tokens=4,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=256,num_output_tokens=64,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=256,num_output_tokens=8,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=512,num_output_tokens=1,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=512,num_output_tokens=16,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=512,num_output_tokens=2,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=512,num_output_tokens=32,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=512,num_output_tokens=4,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=512,num_output_tokens=64,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_ul2,tokenizer=google_ul2,num_prompt_tokens=512,num_output_tokens=8,stop=hash,global_prefix=nlg | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1024,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1024,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1024,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1024,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1024,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1024,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1024,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1536,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1536,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1536,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1536,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1536,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1536,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=1536,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=256,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=256,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=256,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=256,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=256,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=256,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=256,num_output_tokens=8 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=512,num_output_tokens=1 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=512,num_output_tokens=16 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=512,num_output_tokens=2 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=512,num_output_tokens=32 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=512,num_output_tokens=4 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=512,num_output_tokens=64 | generation |
synthetic_efficiency:random=None,model=together_yalm,tokenizer=yandex_yalm,num_prompt_tokens=512,num_output_tokens=8 | generation |
synthetic_reasoning:mode=induction,model=AlephAlpha_luminous-base | generation |
synthetic_reasoning:mode=induction,model=AlephAlpha_luminous-extended | generation |
synthetic_reasoning:mode=induction,model=AlephAlpha_luminous-supreme | generation |
synthetic_reasoning:mode=induction,model=ai21_j1-grande | generation |
synthetic_reasoning:mode=induction,model=ai21_j1-grande-v2-beta | generation |
synthetic_reasoning:mode=induction,model=ai21_j1-jumbo | generation |
synthetic_reasoning:mode=induction,model=ai21_j1-large | generation |
synthetic_reasoning:mode=induction,model=ai21_j2-grande | generation |
synthetic_reasoning:mode=induction,model=ai21_j2-jumbo | generation |
synthetic_reasoning:mode=induction,model=ai21_j2-large | generation |
synthetic_reasoning:mode=induction,model=anthropic_stanford-online-all-v4-s3 | generation |
synthetic_reasoning:mode=induction,model=cohere_command-medium-beta | generation |
synthetic_reasoning:mode=induction,model=cohere_command-xlarge-beta | generation |
synthetic_reasoning:mode=induction,model=cohere_large-20220720 | generation |
synthetic_reasoning:mode=induction,model=cohere_medium-20220720 | generation |
synthetic_reasoning:mode=induction,model=cohere_medium-20221108 | generation |
synthetic_reasoning:mode=induction,model=cohere_small-20220720 | generation |
synthetic_reasoning:mode=induction,model=cohere_xlarge-20220609 | generation |
synthetic_reasoning:mode=induction,model=cohere_xlarge-20221108 | generation |
synthetic_reasoning:mode=induction,model=microsoft_TNLGv2_530B | generation |
synthetic_reasoning:mode=induction,model=microsoft_TNLGv2_7B | generation |
synthetic_reasoning:mode=induction,model=openai_ada | generation |
synthetic_reasoning:mode=induction,model=openai_babbage | generation |
synthetic_reasoning:mode=induction,model=openai_code-cushman-001 | generation |
synthetic_reasoning:mode=induction,model=openai_code-davinci-002 | generation |
synthetic_reasoning:mode=induction,model=openai_curie | generation |
synthetic_reasoning:mode=induction,model=openai_davinci | generation |
synthetic_reasoning:mode=induction,model=openai_gpt-3.5-turbo-0301 | generation |
synthetic_reasoning:mode=induction,model=openai_text-ada-001 | generation |
synthetic_reasoning:mode=induction,model=openai_text-babbage-001 | generation |
synthetic_reasoning:mode=induction,model=openai_text-curie-001 | generation |
synthetic_reasoning:mode=induction,model=openai_text-davinci-002 | generation |
synthetic_reasoning:mode=induction,model=openai_text-davinci-003 | generation |
synthetic_reasoning:mode=induction,model=together_bloom | generation |
synthetic_reasoning:mode=induction,model=together_glm,stop=hash | generation |
synthetic_reasoning:mode=induction,model=together_gpt-j-6b | generation |
synthetic_reasoning:mode=induction,model=together_gpt-neox-20b | generation |
synthetic_reasoning:mode=induction,model=together_opt-175b | generation |
synthetic_reasoning:mode=induction,model=together_opt-66b | generation |
synthetic_reasoning:mode=induction,model=together_redpajama-incite-base-3b-v1 | generation |
synthetic_reasoning:mode=induction,model=together_t0pp,stop=hash | generation |
synthetic_reasoning:mode=induction,model=together_t5-11b,stop=hash | generation |
synthetic_reasoning:mode=induction,model=together_ul2,stop=hash,global_prefix=nlg | generation |
synthetic_reasoning:mode=induction,model=together_yalm | generation |
synthetic_reasoning:mode=induction,model=writer_palmyra-instruct-30 | generation |
synthetic_reasoning:mode=induction,model=writer_palmyra-x | generation |
synthetic_reasoning:mode=pattern_match,model=AlephAlpha_luminous-base | generation |
synthetic_reasoning:mode=pattern_match,model=AlephAlpha_luminous-extended | generation |
synthetic_reasoning:mode=pattern_match,model=AlephAlpha_luminous-supreme | generation |
synthetic_reasoning:mode=pattern_match,model=ai21_j1-grande | generation |
synthetic_reasoning:mode=pattern_match,model=ai21_j1-grande-v2-beta | generation |
synthetic_reasoning:mode=pattern_match,model=ai21_j1-jumbo | generation |
synthetic_reasoning:mode=pattern_match,model=ai21_j1-large | generation |
synthetic_reasoning:mode=pattern_match,model=ai21_j2-grande | generation |
synthetic_reasoning:mode=pattern_match,model=ai21_j2-jumbo | generation |
synthetic_reasoning:mode=pattern_match,model=ai21_j2-large | generation |
synthetic_reasoning:mode=pattern_match,model=anthropic_stanford-online-all-v4-s3 | generation |
synthetic_reasoning:mode=pattern_match,model=cohere_command-medium-beta | generation |
synthetic_reasoning:mode=pattern_match,model=cohere_command-xlarge-beta | generation |
synthetic_reasoning:mode=pattern_match,model=cohere_large-20220720 | generation |
synthetic_reasoning:mode=pattern_match,model=cohere_medium-20220720 | generation |
synthetic_reasoning:mode=pattern_match,model=cohere_medium-20221108 | generation |
synthetic_reasoning:mode=pattern_match,model=cohere_small-20220720 | generation |
synthetic_reasoning:mode=pattern_match,model=cohere_xlarge-20220609 | generation |
synthetic_reasoning:mode=pattern_match,model=cohere_xlarge-20221108 | generation |
synthetic_reasoning:mode=pattern_match,model=microsoft_TNLGv2_530B | generation |
synthetic_reasoning:mode=pattern_match,model=microsoft_TNLGv2_7B | generation |
synthetic_reasoning:mode=pattern_match,model=openai_ada | generation |
synthetic_reasoning:mode=pattern_match,model=openai_babbage | generation |
synthetic_reasoning:mode=pattern_match,model=openai_code-cushman-001 | generation |
synthetic_reasoning:mode=pattern_match,model=openai_code-davinci-002 | generation |
synthetic_reasoning:mode=pattern_match,model=openai_curie | generation |
synthetic_reasoning:mode=pattern_match,model=openai_davinci | generation |
synthetic_reasoning:mode=pattern_match,model=openai_gpt-3.5-turbo-0301 | generation |
synthetic_reasoning:mode=pattern_match,model=openai_text-ada-001 | generation |
synthetic_reasoning:mode=pattern_match,model=openai_text-babbage-001 | generation |
synthetic_reasoning:mode=pattern_match,model=openai_text-curie-001 | generation |
synthetic_reasoning:mode=pattern_match,model=openai_text-davinci-002 | generation |
synthetic_reasoning:mode=pattern_match,model=openai_text-davinci-003 | generation |
synthetic_reasoning:mode=pattern_match,model=together_bloom | generation |
synthetic_reasoning:mode=pattern_match,model=together_glm,stop=hash | generation |
synthetic_reasoning:mode=pattern_match,model=together_gpt-j-6b | generation |
synthetic_reasoning:mode=pattern_match,model=together_gpt-neox-20b | generation |
synthetic_reasoning:mode=pattern_match,model=together_opt-175b | generation |
synthetic_reasoning:mode=pattern_match,model=together_opt-66b | generation |
synthetic_reasoning:mode=pattern_match,model=together_redpajama-incite-base-3b-v1 | generation |
synthetic_reasoning:mode=pattern_match,model=together_t0pp,stop=hash | generation |
synthetic_reasoning:mode=pattern_match,model=together_t5-11b,stop=hash | generation |
synthetic_reasoning:mode=pattern_match,model=together_ul2,stop=hash,global_prefix=nlg | generation |
synthetic_reasoning:mode=pattern_match,model=together_yalm | generation |
synthetic_reasoning:mode=pattern_match,model=writer_palmyra-instruct-30 | generation |
synthetic_reasoning:mode=pattern_match,model=writer_palmyra-x | generation |
synthetic_reasoning:mode=variable_substitution,model=AlephAlpha_luminous-base | generation |
synthetic_reasoning:mode=variable_substitution,model=AlephAlpha_luminous-extended | generation |
synthetic_reasoning:mode=variable_substitution,model=AlephAlpha_luminous-supreme | generation |
synthetic_reasoning:mode=variable_substitution,model=ai21_j1-grande | generation |
synthetic_reasoning:mode=variable_substitution,model=ai21_j1-grande-v2-beta | generation |
synthetic_reasoning:mode=variable_substitution,model=ai21_j1-jumbo | generation |
synthetic_reasoning:mode=variable_substitution,model=ai21_j1-large | generation |
synthetic_reasoning:mode=variable_substitution,model=ai21_j2-grande | generation |
synthetic_reasoning:mode=variable_substitution,model=ai21_j2-jumbo | generation |
synthetic_reasoning:mode=variable_substitution,model=ai21_j2-large | generation |
synthetic_reasoning:mode=variable_substitution,model=anthropic_stanford-online-all-v4-s3 | generation |
synthetic_reasoning:mode=variable_substitution,model=cohere_command-medium-beta | generation |
synthetic_reasoning:mode=variable_substitution,model=cohere_command-xlarge-beta | generation |