Prabhas2002 commited on
Commit
fb59929
1 Parent(s): 7e88654

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -15
README.md CHANGED
@@ -1,24 +1,57 @@
1
  ---
2
  library_name: peft
 
 
 
 
 
 
 
3
  ---
4
- ## Training procedure
5
 
 
6
 
7
- The following `bitsandbytes` quantization config was used during training:
8
- - quant_method: bitsandbytes
9
- - load_in_8bit: True
10
- - load_in_4bit: True
11
- - llm_int8_threshold: 6.0
12
- - llm_int8_skip_modules: None
13
- - llm_int8_enable_fp32_cpu_offload: False
14
- - llm_int8_has_fp16_weight: False
15
- - bnb_4bit_quant_type: fp4
16
- - bnb_4bit_use_double_quant: False
17
- - bnb_4bit_compute_dtype: float32
18
 
 
19
 
20
- ### Framework versions
 
 
 
 
 
 
 
21
 
22
- - PEFT 0.4.0
23
 
24
- - PEFT 0.4.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  library_name: peft
3
+ license: mit
4
+ datasets:
5
+ - uonlp/CulturaX
6
+ - ai4bharat/samanantar
7
+ language:
8
+ - te
9
+ pipeline_tag: text2text-generation
10
  ---
11
+ ## Telugu LLaMA 7B Base Model for Causal LM(v1.0)
12
 
13
+ ### Overview
14
 
15
+ Welcome to the release of the Telugu LLaMA 7B base model – a significant step forward in Language Learning Models (LLMs) for Telugu. This model is specifically designed for Causal Language Modeling (LM) tasks and is ready for immediate inference. It can also be fine-tuned for more specialized Natural Language Processing (NLP) applications.
 
 
 
 
 
 
 
 
 
 
16
 
17
+ ### Key Features
18
 
19
+ - **Model Type:** 7B parameter model for Causal LM
20
+ - **Language:** Telugu
21
+ - **Vocabulary Size:** 50k tokens (32k original + 18k new added)
22
+ - **Training Data:** Smanathar Dataset (60k samples)
23
+ - **Source Model:** Llama7b-chat-hf
24
+ - **Training Precision:** float16
25
+ - **License:** MIT
26
+ - **Code:** https://github.com/prabhas2002/ENGTOTEL-Transalatron/tree/main/Llama2-7b
27
 
28
+ ### Model Performance
29
 
30
+ - **Causal Language Modeling:** Generates fluent and contextually relevant Telugu text.
31
+ - **Fine-Tuning:** Primed for further fine-tuning on specific Telugu NLP tasks.
32
+ - **Multilingual Capability:** Capable of handling Telugu and potentially other languages.
33
+
34
+
35
+ #### Hugging Face Model Hub
36
+
37
+ - **Model Download:** Available on Hugging Face's model hub for download and offline use.
38
+ - **Model Pipelines:** Utilize through Hugging Face's pipelines for text generation and understanding tasks.
39
+ - **Fine-Tuning:** Customize the model for your specific Telugu NLP tasks by fine-tuning on relevant datasets.
40
+
41
+ ### Citation
42
+
43
+ If you use this Telugu LLaMA 7B base model in your work, please cite it using the following BibTeX entry:
44
+
45
+ ```bibtex
46
+ @article{PreTrained_Telugu_Llama7b,
47
+ title={Telugu LLaMA 7B Base Model for Causal LM},
48
+ author={Onteru Prabhas Reddy},
49
+ journal={Hugging Face Model Hub},
50
+ year={2024},
51
+ url=https://huggingface.co/Prabhas2002/PreTrained_Telugu_Llama7b
52
+ }
53
+ ```
54
+
55
+ ### License Information
56
+
57
+ Please refer to the license information provided with the model for details on usage and distribution.