H4rmoniousPampero / README.md
neovalle's picture
Update README.md
63dcf26
|
raw
history blame
No virus
1.64 kB
---
tags:
- autotrain
- text-generation
widget:
- text: 'Tell me about bees.'
library_name: transformers
pipeline_tag: text-generation
---
# Model Card for Model neovalle/H4rmoniousPampero
## Model Details
### Model Description
This is model is a version of HuggingFaceH4/zephyr-7b-alpha fine-tuned with the H4rmony dataset, which aims
to better align the model with ecological values through the use of ecolinguistics principles.
- **Developed by:** Jorge Vallego
- **Funded by :** Neovalle Ltd.
- **Shared by :** airesearch@neovalle.co.uk
- **Model type:** mistral
- **Language(s) (NLP):** Primarily English
- **License:** MIT
- **Finetuned from model:** HuggingFaceH4/zephyr-7b-alpha
## Uses
Intended as PoC to show the effect of H4rmony dataset.
### Direct Use
For testing purposes to gain insight in order to help with the continous improvement of the H4rmony dataset.
### Downstream Use
Its direct use in applications is not recommended as this model is under testing for a specific task only
### Out-of-Scope Use
Not meant to be used other than testing and evaluation of the H4rmony dataset.
## Bias, Risks, and Limitations
This model might produce biased completions already existing in the base model and unintentionally introduced during fine-tuning.
## How to Get Started with the Model
It can be loaded and run in a free Colab instance.
Code to load base and finetuned models to compare outputs:
https://github.com/Neovalle/H4rmony/blob/main/H4rmoniousPampero.ipynb
## Training Details
Autotrained reward model
### Training Data
H4rmony Dataset - https://huggingface.co/datasets/neovalle/H4rmony