From bab9e8e20e71e526175cb3d93752cefbcbb9e6d2 Mon Sep 17 00:00:00 2001 From: DesmonDay <908660116@qq.com> Date: Mon, 30 Oct 2023 14:55:27 +0800 Subject: [PATCH] rm model_type --- llm/run_pretrain.py | 1 - scripts/regression/requirements_ci.txt | 3 ++- tests/fixtures/llm/pretrain.yaml | 3 --- tests/requirements.txt | 1 - 4 files changed, 2 insertions(+), 6 deletions(-) diff --git a/llm/run_pretrain.py b/llm/run_pretrain.py index b0c199f1fd8a..d684e0245cbd 100644 --- a/llm/run_pretrain.py +++ b/llm/run_pretrain.py @@ -115,7 +115,6 @@ class ModelArguments: Arguments pertaining to which model/config/tokenizer we are going to pre-train from. """ - model_type: Optional[str] = field(default="llama", metadata={"help": "Use for CI test."}) model_name_or_path: str = field( default="__internal_testing__/tiny-random-llama", metadata={ diff --git a/scripts/regression/requirements_ci.txt b/scripts/regression/requirements_ci.txt index caae5fce67fe..8d61a9f60b4e 100644 --- a/scripts/regression/requirements_ci.txt +++ b/scripts/regression/requirements_ci.txt @@ -40,4 +40,5 @@ onnxruntime sacremoses soundfile librosa -gradio \ No newline at end of file +gradio +tiktoken diff --git a/tests/fixtures/llm/pretrain.yaml b/tests/fixtures/llm/pretrain.yaml index 1b4046355fd0..8197d1bc6fcf 100644 --- a/tests/fixtures/llm/pretrain.yaml +++ b/tests/fixtures/llm/pretrain.yaml @@ -20,13 +20,10 @@ pretrain: continue_training: 1 default: llama: - model_type: llama model_name_or_path: __internal_testing__/tiny-random-llama chatglm: - model_type: chatglm model_name_or_path: __internal_testing__/tiny-fused-chatglm qwen: - model_type: qwen model_name_or_path: __internal_testing__/tiny-fused-qwen inference-predict: diff --git a/tests/requirements.txt b/tests/requirements.txt index 5191a64e0dab..000a843debf5 100644 --- a/tests/requirements.txt +++ b/tests/requirements.txt @@ -7,4 +7,3 @@ tool_helpers fast_tokenizer_python sacremoses pydantic==1.10.9 -tiktoken