Skip to content

Commit

Permalink
#0: Minor fixes
Browse files Browse the repository at this point in the history
  • Loading branch information
mtairum committed Dec 3, 2024
1 parent e33367e commit e8d8e39
Show file tree
Hide file tree
Showing 3 changed files with 7 additions and 8 deletions.
2 changes: 1 addition & 1 deletion models/demos/llama3/demo/demo.py
Original file line number Diff line number Diff line change
Expand Up @@ -920,5 +920,5 @@ def test_llama_demo(
sampling_params=sampling_params,
instruct_mode=instruct,
is_ci_env=is_ci_env,
print_to_file=True, # TODO Miguel set this to false before merging
print_to_file=False,
)
9 changes: 4 additions & 5 deletions models/demos/llama3/lt
Original file line number Diff line number Diff line change
Expand Up @@ -729,13 +729,11 @@ def run_entry_command(entry, screen_lock, output_entries, screen_needs_update):
# Open log file
entry.open_log_file()

# TODO Miguel: Update LT to the final demo.py configuration

# Define command shortcuts
command_shortcuts = {
"demo": "pytest models/demos/llama3/demo/demo.py -k 'instruct-1 and attention-1'",
"demo-batch32": "pytest models/demos/llama3/demo/demo.py -k 'instruct-1 and attention-32'",
"demo-1layer": "pytest models/demos/llama3/demo/demo.py -k single_layer",
"demo": "pytest models/demos/llama3/demo/demo.py -k performance-batch-1",
"demo-32": "pytest models/demos/llama3/demo/demo.py -k performance-batch-32",
"demo-long": "pytest models/demos/llama3/demo/demo.py -k long",
"attention": "pytest models/demos/llama3/tests/test_llama_attention.py",
"attention-prefill": "pytest models/demos/llama3/tests/test_llama_attention_prefill.py",
"mlp": "pytest models/demos/llama3/tests/test_llama_mlp.py",
Expand All @@ -747,6 +745,7 @@ def run_entry_command(entry, screen_lock, output_entries, screen_needs_update):
"model": "pytest models/demos/llama3/tests/test_llama_model.py -k 'performance-128 and full'",
"model-quick": "pytest models/demos/llama3/tests/test_llama_model.py -k 'performance-128 and quick'",
"model-prefill": "pytest models/demos/llama3/tests/test_llama_model_prefill.py -k performance-4096",
# Vision tests (require 11B weights)
"vision-mlp": "pytest models/demos/llama3/tests/multimodal/test_llama_image_mlp.py",
"vision-attn": "pytest models/demos/llama3/tests/multimodal/test_llama_image_attention.py",
"vision-block": "pytest models/demos/llama3/tests/multimodal/test_llama_image_block.py",
Expand Down
4 changes: 2 additions & 2 deletions models/demos/llama3/tests/test_llama_attention.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,11 +36,11 @@
"paged_attention",
(
True,
# False,
False,
),
ids=(
"paged_attention",
# "default_attention",
"default_attention",
),
)
@pytest.mark.parametrize(
Expand Down

0 comments on commit e8d8e39

Please sign in to comment.