File size: 533 Bytes
6f1e2aa
 
e34dd4f
 
 
 
6f1e2aa
e34dd4f
 
127323e
 
 
 
e34dd4f
127323e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
---
license: apache-2.0
datasets:
- berkeley-nest/Nectar
language:
- en
---
# Use cases
This model is used to deep clean the Rhino dataset, making it a higher quality dataset. This model achieved an average MSE loss of 0.095 during training.
We recommend to use the sigmoid function to turn the logits into probabilities:
```python
1 / (1 + torch.exp(logits))
```
# Training
Using trl's RewardTrainer, this model was trained on berkeley-nest/Nectar. The dataset is curated on-the-fly during training, as explained in the Rhino repo.