File size: 2,826 Bytes
bb436de 7ecf5ef 713ae89 f83c3ff 7ecf5ef 713ae89 7ecf5ef 157ecd6 7ecf5ef 157ecd6 7ecf5ef 713ae89 6f6d6d3 713ae89 7ecf5ef 713ae89 7ecf5ef 157ecd6 7ecf5ef 713ae89 7ecf5ef 713ae89 7ecf5ef 713ae89 7ecf5ef 157ecd6 7ecf5ef 157ecd6 713ae89 6634bc1 713ae89 6f6d6d3 713ae89 8bba3c0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 |
---
license: apache-2.0
datasets:
- nvidia/HelpSteer2
- flytech/python-codes-25k
language:
- en
base_model:
- unsloth/Llama-3.2-3B-Instruct
tags:
- conversational
- chatbot
- ezcz
- code
- reasoning
---
base_model: unsloth/Llama-3.2-3B-instruct
library_name: peft
tags:
- chatbot
- code-generation
- reasoning
- instruct-tuning
datasets:
- nvidia/HelpSteer2
- flytech/python-codes-25k
license: apache-2.0
---
# Bright Llama 3B Chat
Bright Llama is a fine-tuned AI model designed to further excel at coding and reasoning tasks without sacrificing much of the generality of the base model. It is based on `unsloth/Llama-3.2-3B-instruct` and fine-tuned using specialized datasets for enhancing its problem-solving and code-generation capabilities. The model is optimized for generating Python code and providing logical explanations.
### Model Description
Bright Llama is developed to assist users with technical and problem-solving tasks, focusing on generating code and providing reasoned responses. The fine-tuning datasets, `nvidia/HelpSteer2` and `flytech/python-codes-25k`, were specifically chosen to enhance its performance in both reasoning, and coding-related scenarios.
- **Developed by:** ezcz
- **Model type:** Llama-based model
- **Languages:** English
- **License:** Apache-2.0
- **Fine-tuned from:** `unsloth/Llama-3.2-3B-instruct`
### Model Sources
- **Repository:** [ezcz/bright-llama-3b-chat](https://huggingface.co/ezcz/bright-llama-3b-chat)
## Uses
### Direct Use
- Generating Python code snippets
- Creative writing
- Assisting with tasks that require reasoning
### Downstream Use
The model can be further fine-tuned for domain-specific technical support, coding tutorials, or reasoning-based applications.
### Out-of-Scope Use
- Not designed for generating explicit, harmful, or unethical content.
## Bias, Risks, and Limitations
Bright Llama focuses on coding and reasoning, but may still exhibit biases present in the training data. It is recommended to avoid deploying it in sensitive contexts without additional evaluation.
### Recommendations
- For best results, use Bright Llama for coding, creative writing, and reasoning tasks.
## How to Get Started with the Model
Here’s an example of how to use Bright Llama for coding and reasoning tasks:
```python
import torch
from transformers import pipeline
model_id = "ezcz/Llama-3.2-3B-BrightLlamaChat-LoRA"
pipe = pipeline(
"text-generation",
model=model_id,
torch_dtype=torch.float16,
device_map="auto",
)
messages = [
{"role": "system", "content": ""},
{"role": "user", "content": "What is the number that rhymes with the word for a large plant?"},
]
outputs = pipe(
messages,
max_new_tokens=256,
temperature=0.1,
top_p=0.9,
top_k=60,
)
print(outputs[0]["generated_text"][-1]) |