diff --git a/examples/mistral/Mistral-7b-example/README.md b/examples/mistral/Mistral-7b-example/README.md index 1d5e03f3d4..2d5ac87a17 100644 --- a/examples/mistral/Mistral-7b-example/README.md +++ b/examples/mistral/Mistral-7b-example/README.md @@ -10,6 +10,3 @@ Hardware Utilized: For reference, the fine-tuning in this guide was performed us **Uploading to HuggingFace 🤗:** To upload your fine-tuned model to Hugging Face, include the following files: ![Screenshot 2024-01-19 213932](https://github.com/OpenAccess-AI-Collective/axolotl/assets/138583191/d660eb84-2d76-46a1-9846-cf0aeb3006d9) - - -By Tilemachos Chatzipapas t.chatzipapas@gmail.com diff --git a/examples/mistral/Mistral-7b-example/config.yml b/examples/mistral/Mistral-7b-example/config.yml index b103b447e8..84be18d15e 100644 --- a/examples/mistral/Mistral-7b-example/config.yml +++ b/examples/mistral/Mistral-7b-example/config.yml @@ -11,7 +11,7 @@ strict: false datasets: - path: tilemachos/Demo-Dataset #Path to json dataset file in huggingface #for type,conversation arguments read axolotl readme and pick what is suited for your project, I wanted a chatbot and put sharegpt and chatml - type: sharegpt + type: sharegpt conversation: chatml dataset_prepared_path: tilemachos/Demo-Dataset #Path to json dataset file in huggingface val_set_size: 0.05 @@ -25,7 +25,7 @@ lora_dropout: 0.05 lora_target_modules: - q_proj - v_proj - + sequence_len: 512 sample_packing: false pad_to_sequence_len: true