File size: 1,574 Bytes
35da623
 
 
 
 
 
 
 
 
 
860a8a8
35da623
 
 
4987a8f
36ccb1a
 
 
4987a8f
35da623
 
4987a8f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
---
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- gemma
- trl
- sft
base_model: google/gemma-7b
---

# Tamil Large Language Model

Developing a large language model for Tamil is a significant achievement, highlighting the potential for advanced AI technologies to support and enrich Tamil digital content. Created by Mervin Praison, this model stands as a testament to innovation and dedication towards enhancing language processing capabilities for Tamil speakers around the globe. It is designed to understand, interpret, and generate Tamil text, offering invaluable tools for businesses, educators, and researchers. By leveraging this model, we can unlock new opportunities for content creation, automate and improve customer support, and contribute to preserving and promoting the Tamil language in the digital age.

## Uploaded  model

- **Developed by:** mervinpraison
- **Instruction Fine Tuned :** alpaca tamil dataset

## How to use?

```
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("mervinpraison/tamil-large-language-model-7b-v1.0")
model = AutoModelForCausalLM.from_pretrained("mervinpraison/tamil-large-language-model-7b-v1.0")

query_to_llm = "ஆரோக்கியமாக இருப்பதற்கான இரண்டு வழிகள்"
inputs = tokenizer.encode(query_to_llm, return_tensors="pt")
outputs = model.generate(inputs, max_length=200)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
```