|
--- |
|
license: apache-2.0 |
|
--- |
|
# β¨ Orion-7B |
|
**Orion is based on llama-7B trained on the latest deep learning papers that were published in notable conferences with various PEFT methods. It is made available under the Apache 2.0 license.** |
|
*Paper coming soon π.* |
|
|
|
|
|
## What is so special about Orion? |
|
* **Comprehensive performance and evaluation analysis of PEFT methods for knowledge editing** |
|
* **Novel prompting framekwork for knowledge editing dataset generation** |
|
* **Large knowledge editing dataset from recent deep learning papers published in top conferences** |
|
* |
|
β οΈ **This is a raw, fine-tuned model, which should be further instruction-tuned for production level performance.** |
|
|
|
π₯ **Orion LLM require PyTorch 2.0** |
|
|
|
You will need **at least 20-25GB of memory** to swiftly run inference with Orion-7B. |
|
|
|
## Model Details |
|
|
|
### Model Description |
|
- **Developed by:** [AttentionX](https://atttentionx.github.io); |
|
- **Base model:** llama-7B; |
|
- **Training method: LoRA, llama-adapter, llama-adapterV2, full fine-tuning** |
|
- **Language(s) (NLP):** English; |
|
- **License:** Apache 2.0 license. |
|
|
|
### Model Source |
|
- **Paper:** *coming soon*. |
|
|
|
|
|
## Training Details |
|
*Paper coming soon π.* |
|
|
|
|
|
## Evaluation |
|
*Paper coming soon.* |
|
|
|
|
|
## Hardware |
|
Orion was trained on a single a100 |
|
|
|
## Citation |
|
*Paper coming soon* π. In the meanwhile, you can use the following information to cite: |
|
``` |
|
@article{orion7b, |
|
title={{Orion-7B}: knolwedge editing via peft}, |
|
author={Adam Lee, Sungkyung Kim, Junyoung Park, Sunho Jung, Jusang Oh}, |
|
year={2023} |
|
} |
|
``` |
|
|
|
|
|
## License |
|
|
|
Orion is made available under the Apache 2.0 license. |
|
|
|
## Contact |
|
attentionx.ai@gmail.com |