osmanorhan commited on
Commit
7a7ce0e
1 Parent(s): 10a3311

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -3
README.md CHANGED
@@ -1,3 +1,33 @@
1
- ---
2
- license: llama3.1
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3.1
3
+ language:
4
+ - tr
5
+ ---
6
+
7
+ This is a quantized version of the BrewInteractive/fikri-3.1-8B-Instruct model.
8
+
9
+ * Original model: fikri-3.1-8B-Instruct
10
+ * Base model: LLaMA-3.1-8B
11
+ * Quantization: Q4_K_M
12
+
13
+ * Optimized for faster inference and reduced memory usage while maintaining performance
14
+ * Built on the LLaMA 3.1 architecture (8B)
15
+ * Fine-tuned for Turkish language tasks
16
+ * Quantized for improved efficiency
17
+
18
+ # How to use
19
+
20
+ 1. Install llama.cpp:
21
+ * For macOS, use Homebrew:
22
+ ```
23
+ brew install llama.cpp
24
+ ```
25
+ * For other operating systems, follow the installation instructions on the [llama.cpp GitHub repository](https://github.com/ggerganov/llama.cpp).
26
+
27
+ 2. Download the quantized GGUF file from this repository's Files section.
28
+
29
+ 3. Run the following command for conversation mode:
30
+
31
+ ```
32
+ llama-cli -m ./fikri-3.1-8B-Instruct-Q4_K_M.gguf --no-mmap -fa -c 4096 --temp 0.8 -if --in-prefix "<|start_header_id|>user<|end_header_id|>\n\n" --in-suffix "<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
33
+ ```