szymonrucinski commited on
Commit
06bc2b9
1 Parent(s): f15815f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -56,5 +56,6 @@ The model was trained using AdamW optimizer, with specific hyperparameters aimed
56
  ## Conclusions
57
  The `Curie-7B-v1` model, through LAPT, matches foundational models on eight downstream tasks with significantly less data. Its versatility in generating Polish text and the ability to be transformed into classifiers, regressors, and AI assistants highlights the method's effectiveness. This open-source Polish LLM provides a foundation for developing efficient business solutions.
58
 
59
- ## Citation
60
- If you use `Curie-7B-v1` in your research, please cite:
 
 
56
  ## Conclusions
57
  The `Curie-7B-v1` model, through LAPT, matches foundational models on eight downstream tasks with significantly less data. Its versatility in generating Polish text and the ability to be transformed into classifiers, regressors, and AI assistants highlights the method's effectiveness. This open-source Polish LLM provides a foundation for developing efficient business solutions.
58
 
59
+ ## Research Paper
60
+ Work and details regarding this model are described in the reserach paper [Efficient Language Adaptive Pre-training: Extending State-of-the-Art Large Language Models for Polish](https://arxiv.org/abs/2402.09759) by Szymon Ruciński.
61
+