From 77844936cce1e99d7760afe782c47ca15637d2dc Mon Sep 17 00:00:00 2001 From: camenduru <54370274+camenduru@users.noreply.github.com> Date: Fri, 13 Oct 2023 23:05:36 +0300 Subject: [PATCH] test --- LLaVA_13b_8bit_colab.ipynb | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/LLaVA_13b_8bit_colab.ipynb b/LLaVA_13b_8bit_colab.ipynb index 7eedae3..06bc21c 100644 --- a/LLaVA_13b_8bit_colab.ipynb +++ b/LLaVA_13b_8bit_colab.ipynb @@ -84,8 +84,6 @@ "import torch\n", "\n", "model_path = \"4bit/llava-v1.5-13b-5GB\"\n", - "tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False)\n", - "\n", "# model_path = \"4bit/llava-v1.5-13b-4GB-8bit\"\n", "# model = LlavaLlamaForCausalLM.from_pretrained(model_path)\n", "# model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, load_in_8bit=True, device_map=\"auto\")\n", @@ -98,7 +96,8 @@ " bnb_4bit_use_double_quant=True,\n", " bnb_4bit_quant_type='nf4'\n", ")\n", - "model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs)" + "model = LlavaLlamaForCausalLM.from_pretrained(model_path, low_cpu_mem_usage=True, **kwargs)\n", + "tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False)" ] }, {