|
--- |
|
license: mit |
|
language: |
|
- en |
|
--- |
|
|
|
## Model Details |
|
- *Finetuned+Capable for laptop |
|
### Model Description |
|
------------ |
|
Capable for run in Low-end **laptop** |
|
|
|
- **Developed by:** [Tiny-llama]("https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0/tree/main") |
|
|
|
- **Finetuned from model [optional]:** [Tiny-llama]("https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0/tree/main") |
|
|
|
|
|
|
|
## Uses |
|
```python |
|
from llama_cpp import Llama |
|
|
|
llm = Llama( |
|
model_path="path/to/llama", |
|
# n_gpu_layers=-1, # Uncomment to use GPU acceleration |
|
# seed=1337, # Uncomment to set a specific seed |
|
# n_ctx=2048, # Uncomment to increase the context window |
|
) |
|
output = llm( |
|
"Q: Name the planets in the solar system? A: ", # Prompt |
|
max_tokens=32, # Generate up to 32 tokens, set to None to generate up to the end of the context window |
|
stop=["Q:", "\n"], # Stop generating just before the model would generate a new question |
|
echo=True # Echo the prompt back in the output |
|
) # Generate a completion, can also call create_completion |
|
print(output) |
|
``` |
|
|