Update README.md
Browse files
README.md
CHANGED
@@ -41,8 +41,9 @@ The following quantized versions are available:
|
|
41 |
./server -m models/FineLlama-3.2-3B-Instruct-ead-Q4_K_M.gguf -c 4096
|
42 |
```
|
43 |
|
|
|
|
|
44 |
```python
|
45 |
-
# Example using llama-cpp-python library
|
46 |
from llama_cpp import Llama
|
47 |
query = "..."
|
48 |
llm = Llama.from_pretrained(
|
@@ -62,6 +63,11 @@ output = llm.create_chat_completion(
|
|
62 |
)
|
63 |
print(output["choices"][0]["message"]["content"])
|
64 |
```
|
|
|
|
|
|
|
|
|
|
|
65 |
|
66 |
## Quantization Details
|
67 |
|
|
|
41 |
./server -m models/FineLlama-3.2-3B-Instruct-ead-Q4_K_M.gguf -c 4096
|
42 |
```
|
43 |
|
44 |
+
### Example using llama-cpp-python library
|
45 |
+
|
46 |
```python
|
|
|
47 |
from llama_cpp import Llama
|
48 |
query = "..."
|
49 |
llm = Llama.from_pretrained(
|
|
|
63 |
)
|
64 |
print(output["choices"][0]["message"]["content"])
|
65 |
```
|
66 |
+
### Example using Ollama
|
67 |
+
|
68 |
+
```bash
|
69 |
+
ollama run hf.co/Geraldine/FineLlama-3.2-3B-Instruct-ead-GGUF:Q4_K_M
|
70 |
+
```
|
71 |
|
72 |
## Quantization Details
|
73 |
|