diff --git a/models/demos/llama3/tests/test_llama_perf.py b/models/demos/llama3/tests/test_llama_perf.py index 260a0221e9da..ea3c922646db 100644 --- a/models/demos/llama3/tests/test_llama_perf.py +++ b/models/demos/llama3/tests/test_llama_perf.py @@ -52,7 +52,7 @@ def test_llama_model_perf(mesh_device, kv_cache_len, expected_compile_time, use_ elif "3.1-8B" in model_args.DEFAULT_CACHE_PATH: expected_inference_time = 0.07 elif "3.2-11B" in model_args.DEFAULT_CACHE_PATH: - expected_inference_time = 0.07 + expected_inference_time = 0.085 else: assert False, f"Llama model not found. Supported Llama models: [3.2-1B, 3.2-3B, 3.1-8B]"