datalama's picture
Create README.md
fe48447 verified
---
license: other
license_name: exaone
license_link: LICENSE
language:
- en
- ko
tags:
- lg-ai
- exaone
- exaone-3.5
pipeline_tag: text-generation
library_name: transformers
---
# Updates in EXAONE-3.5
## Key Changes
- **RoPE Scaling Parameter**: Added to support longer `context_length`.
- **Memory Optimization**: For the 2.4B model, `tie_word_embeddings` is set to `True` for improved memory efficiency.
⚠️ Using the original [Llamafy script](https://huggingface.co/maywell/EXAONE-3.0-7.8B-Instruct-Llamafied) as-is may lead to performance degradation.
To address this, I have updated the script and uploaded the Llamafied version of the model.
## Special Thanks
- **[@maywell](https://huggingface.co/maywell)**
For updating the code and uploading the model.
- **LG AI Research**
For releasing the original model.
Check out the [original release here](https://huggingface.co/collections/LGAI-EXAONE/exaone-35-674d0e1bb3dcd2ab6f39dbb4).