File size: 1,184 Bytes
9243af0 7984e75 9243af0 7984e75 9243af0 7984e75 9243af0 7984e75 9243af0 7984e75 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 |
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** Two Platforms
- **Model type:** Tokenizer for SUTRA Models. SUTRA Models are dual transformer based Multilingual LLMs
- **Language(s) (NLP):** 50+ Languages including English, Hindi, Gujarati, Bengali, Tamil, Korean, Arabic, Japanese, French, German etc.
- **License:** Proprietary
### Model Sources
- **Paper:** [SUTRA: Scalable Multilingual Language Model Architecture](https://huggingface.co/papers/2405.06694)
- **Demo:** [SUTRA tokenizer comparison](https://huggingface.co/spaces/TWO/sutra-tokenizer-comparison)
## Citation
**BibTeX:**
@misc{bendale2023sutra,
author = {Abhijit Bendale and Michael Sapienza and Steven Ripplinger and Simon Gibbs and Jaewon Lee and Pranav Mistry},
title = {SUTRA: Scalable Multilingual Language Model Architecture},
howpublished = {arXiv preprint arXiv:2405.06694},
year = {2024}
} |