|
--- |
|
library_name: transformers |
|
tags: [bloom-560m, lora] |
|
--- |
|
|
|
# Model Card for Model ID |
|
|
|
This model card describes a transformers model based on the Bloom 560m architecture, fine-tuned with LORA (Linear Regression for Out-of-Distribution Adaptation). This model is intended for advanced users familiar with large language models and LORA. |
|
|
|
### Model Description |
|
|
|
This is a Bloom 560m model fine-tuned with LORA. Bloom 560m is a factual language model from Adept AI, trained on a massive dataset of text and code. LORA is a technique for adapting a pre-trained model to new data without retraining the entire model. |
|
|
|
|
|
- **Developed by:** Tayyib Ul Hassan |
|
<!-- - **Funded by [optional]:** [More Information Needed] |
|
- **Shared by [optional]:** [More Information Needed] --> |
|
- **Model type:** Causal LLM |
|
- **Language(s) (NLP):** English |
|
<!-- - **License:** [More Information Needed] --> |
|
- **Finetuned from model:** Bloom 560m (original model by Adept AI) |
|
|
|
### Model Sources |
|
|
|
<!-- Provide the basic links for the model. --> |
|
|
|
- **Paper:** [Link to paper!](https://arxiv.org/abs/2211.05100) |