From dec9f138147245b30ef995c65e059306e840d789 Mon Sep 17 00:00:00 2001 From: Mishig Date: Fri, 18 Oct 2024 13:28:51 +0200 Subject: [PATCH] [ollama] Add better screenshots (#1457) * [ollama] Add better screenshots * fix imgs --- docs/hub/ollama.md | 24 +++++++++++++++++++++--- 1 file changed, 21 insertions(+), 3 deletions(-) diff --git a/docs/hub/ollama.md b/docs/hub/ollama.md index cd1eea703..48fd1892c 100644 --- a/docs/hub/ollama.md +++ b/docs/hub/ollama.md @@ -6,6 +6,16 @@ Ollama is an application based on llama.cpp to interact with LLMs directly throu Getting started is as simple as: +1. Enable `ollama` under your [Local Apps settings](https://huggingface.co/settings/local-apps). +2. On a model page, choose `ollama` from `Use this model` dropdown. For example: [bartowski/Llama-3.2-1B-Instruct-GGUF](https://huggingface.co/bartowski/Llama-3.2-1B-Instruct-GGUF). + +
+ + +
+ +The snippet would be in format: + ```sh ollama run hf.co/{username}/{repository} ``` @@ -25,14 +35,22 @@ ollama run hf.co/bartowski/Humanish-LLama3-8B-Instruct-GGUF By default, the `Q4_K_M` quantization scheme is used, when it's present inside the model repo. If not, we default to picking one reasonable quant type present inside the repo. -To select a different scheme, simply add a tag: +To select a different scheme, simply: + +1. From `Files and versions` tab on a model page, open GGUF viewer on a particular GGUF file. +2. Choose `ollama` from `Use this model` dropdown. + +
+ + +
+ +The snippet would be in format (quantization tag added): ```sh ollama run hf.co/{username}/{repository}:{quantization} ``` -![guide](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/ollama/guide.png) - For example: ```sh