Skip to content

Commit

Permalink
chore(model gallery): add fusechat-qwen-2.5-7b-instruct (#4380)
Browse files Browse the repository at this point in the history
Signed-off-by: Ettore Di Giacinto <[email protected]>
  • Loading branch information
mudler authored Dec 14, 2024
1 parent 5051074 commit af33483
Showing 1 changed file with 15 additions and 0 deletions.
15 changes: 15 additions & 0 deletions gallery/index.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2018,6 +2018,21 @@
- filename: Evathene-v1.3-Q4_K_M.gguf
sha256: 0f54909b3ddca514994ee16417da8750f56e7bd59581b46ac47625c230e29d1f
uri: huggingface://bartowski/Evathene-v1.3-GGUF/Evathene-v1.3-Q4_K_M.gguf
- !!merge <<: *qwen25
name: "fusechat-qwen-2.5-7b-instruct"
icon: https://huggingface.co/FuseAI/FuseChat-Qwen-2.5-7B-Instruct/resolve/main/FuseChat-3.0.png
urls:
-https://huggingface.co/FuseAI/FuseChat-Qwen-2.5-7B-Instruct
- https://huggingface.co/bartowski/FuseChat-Qwen-2.5-7B-Instruct-GGUF
description: |
We present FuseChat-3.0, a series of models crafted to enhance performance by integrating the strengths of multiple source LLMs into more compact target LLMs. To achieve this fusion, we utilized four powerful source LLMs: Gemma-2-27B-It, Mistral-Large-Instruct-2407, Qwen-2.5-72B-Instruct, and Llama-3.1-70B-Instruct. For the target LLMs, we employed three widely-used smaller models—Llama-3.1-8B-Instruct, Gemma-2-9B-It, and Qwen-2.5-7B-Instruct—along with two even more compact models—Llama-3.2-3B-Instruct and Llama-3.2-1B-Instruct. The implicit model fusion process involves a two-stage training pipeline comprising Supervised Fine-Tuning (SFT) to mitigate distribution discrepancies between target and source LLMs, and Direct Preference Optimization (DPO) for learning preferences from multiple source LLMs. The resulting FuseChat-3.0 models demonstrated substantial improvements in tasks related to general conversation, instruction following, mathematics, and coding. Notably, when Llama-3.1-8B-Instruct served as the target LLM, our fusion approach achieved an average improvement of 6.8 points across 14 benchmarks. Moreover, it showed significant improvements of 37.1 and 30.1 points on instruction-following test sets AlpacaEval-2 and Arena-Hard respectively. We have released the FuseChat-3.0 models on Huggingface, stay tuned for the forthcoming dataset and code.
overrides:
parameters:
model: FuseChat-Qwen-2.5-7B-Instruct-Q4_K_M.gguf
files:
- filename: FuseChat-Qwen-2.5-7B-Instruct-Q4_K_M.gguf
sha256: 8cd8c317769f03125ac753c836ac92c5a76ee0b35502811d0e65bcbb8df9d55c
uri: huggingface://bartowski/FuseChat-Qwen-2.5-7B-Instruct-GGUF/FuseChat-Qwen-2.5-7B-Instruct-Q4_K_M.gguf
- &archfunct
license: apache-2.0
tags:
Expand Down

0 comments on commit af33483

Please sign in to comment.