From 679e15d3d05f200f7984354c0bec8fc27c4b1aeb Mon Sep 17 00:00:00 2001 From: viraatc Date: Tue, 26 Nov 2024 10:00:57 -0800 Subject: [PATCH] correct moe OSL in inference_rules.adoc (#296) Co-authored-by: Pablo Gonzalez --- inference_rules.adoc | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/inference_rules.adoc b/inference_rules.adoc index e576368..2b3de92 100644 --- a/inference_rules.adoc +++ b/inference_rules.adoc @@ -350,7 +350,7 @@ For each of the following benchmarks it is necessary to use the following infere |Summarization (GPT-J) |max_new_tokens |128 | Maximum number of new tokens to generate |Summarization (GPT-J) |early_stopping |True | Use the EOS token to stop generating tokens |Summarization (Llama2) |max_new_tokens |1024 | Maximum number of new tokens to generate -|Text Generation (Mixtral-8x7B) |max_new_tokens |2048 | Maximum number of new tokens to generate +|Text Generation (Mixtral-8x7B) |max_new_tokens |1024 | Maximum number of new tokens to generate |=== == Load Generator