minkdank commited on
Commit
72c077c
·
verified ·
1 Parent(s): 1a39f82

Add README

Browse files
Files changed (1) hide show
  1. README.md +4 -15
README.md CHANGED
@@ -3,7 +3,7 @@ tags:
3
  - gguf
4
  - llama.cpp
5
  - unsloth
6
- - vision-language-model
7
  ---
8
 
9
  # LLAMA-JSON-data-extration - GGUF
@@ -15,18 +15,7 @@ This model was finetuned and converted to GGUF format using [Unsloth](https://gi
15
  - For multimodal models: **llama-mtmd-cli** **-m** model_name.gguf **--mmproj** mmproj_file.gguf
16
 
17
  ## Available Model files:
18
- - `gemma-3-4b-it.Q8_0.gguf`
19
- - `gemma-3-4b-it.BF16-mmproj.gguf`
20
-
21
- ## ⚠️ Ollama Note for Vision Models
22
- **Important:** Ollama currently does not support separate mmproj files for vision models.
23
-
24
- To create an Ollama model from this vision model:
25
- 1. Place the `Modelfile` in the same directory as the finetuned bf16 merged model
26
- 3. Run: `ollama create model_name -f ./Modelfile`
27
- (Replace `model_name` with your desired name)
28
-
29
- This will create a unified bf16 model that Ollama can use.
30
 
31
- ## Note
32
- The model's BOS token behavior was adjusted for GGUF compatibility.
 
3
  - gguf
4
  - llama.cpp
5
  - unsloth
6
+
7
  ---
8
 
9
  # LLAMA-JSON-data-extration - GGUF
 
15
  - For multimodal models: **llama-mtmd-cli** **-m** model_name.gguf **--mmproj** mmproj_file.gguf
16
 
17
  ## Available Model files:
18
+ - `Llama-3.2-3B-Instruct.Q8_0.gguf`
 
 
 
 
 
 
 
 
 
 
 
19
 
20
+ ## Ollama
21
+ An Ollama Modelfile is included for easy deployment.