Transformers
PyTorch
English
trl
rlhf
File size: 2,230 Bytes
cfddde7
 
0452f71
 
cfddde7
 
 
 
 
 
a5343d3
 
7072c92
de16483
7072c92
cfddde7
 
0452f71
 
cfddde7
 
0452f71
cfddde7
0452f71
cfddde7
 
0452f71
 
cfddde7
0452f71
 
cfddde7
0452f71
cfddde7
0452f71
 
 
de16483
0452f71
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
---
license: apache-2.0
language:
- en
tags:
- trl
- transformers
- reinforcement-learning
---

![pull_figure](https://huggingface.co/datasets/trl-internal-testing/example-images/resolve/main/images/stack-llama.png)

# Llama-se-rl-peft
Adapter weights of an RL fine-tuned model based on LLaMa. Authored by Edward Beeching, Younes Belkada, Kashif Rasul, Lewis Tunstall and Leandro von Werra.
For more info check out the [blog post]() and [github example]().


## Model Description
**Llama-se-rl** is a Llama-based model that has been first fine-tuned on the Stack Exchange dataset and then RL fine-tuned using a Stack Exchange Reward Model. This dataset consists of questions and answers from various domains in Stack Exchange, such as programming, mathematics, physics, and more. The model is designed to generate human-like responses to questions in these domains. The model has been training to respond to prompts with the following template:

```
Question: <Query> 

Answer: <Response>
```

## Intended Uses & Limitations
**Llama-se-rl** is intended for use in generating responses to questions related to the Stack Exchange dataset. It is suitable for generating answers to questions in the domains covered by the dataset, such as programming, mathematics, and physics. However, the model may not perform well on questions outside these domains or on questions requiring highly specific or technical knowledge.

## Limitations and Bias
The **Llama-se-rl** model inherits limitations and biases from the Llama model and also those contained in the Stack Exchange dataset. The Stack Exchange dataset may contain biases in terms of the topics it covers and the users who contribute to it. It may not include all possible domains, and the quality of answers may vary. Additionally, the model may generate answers that are incorrect or misleading due to biases in the training data or the inherent limitations of the Llama architecture.

## BibTeX entry and citation info

```bibtex
@misc{beeching2023llama,
  title={StackLLaMa: An RL Fine-tuned LLaMa Model for Stack Exchange Question and Answering},
  author={Beeching, Edward and Belkada, Younes and Rasul, Kashif and Tunstall, Lewis and von Werra, Leandro},
  year={2023}
}
```