File size: 3,295 Bytes
769cfa4
624e3e1
769cfa4
624e3e1
 
 
769cfa4
624e3e1
 
769cfa4
624e3e1
769cfa4
 
 
624e3e1
769cfa4
624e3e1
769cfa4
624e3e1
769cfa4
624e3e1
 
 
 
769cfa4
624e3e1
 
 
769cfa4
624e3e1
 
769cfa4
624e3e1
769cfa4
 
 
 
 
 
 
624e3e1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
769cfa4
624e3e1
769cfa4
624e3e1
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
---
title: Gemma Italian Camoscio Language Model
tags:
- italian-language-generation
- camoscio-dataset
- gemma-2b
- autotrain
datasets:
- camoscio
library_name: transformers
model: theoracle/gemma_italian_camoscio
license: other
---

## Overview

`theoracle/gemma_italian_camoscio` is a cutting-edge model specifically designed for Italian language generation. Leveraging the comprehensive Camoscio dataset, this model enhances the Gemma 2B architecture's capabilities in producing high-quality, contextually accurate Italian text. Developed with AutoTrain, it excels in various Italian text generation tasks, including but not limited to creative writing, article generation, and conversational responses.

## Key Features

- **Italian Language Focus**: Tailored to understand and generate Italian text, capturing the language's nuances and complexities.
- **Camoscio Dataset Training**: Utilizes the rich Camoscio dataset, ensuring the model is well-versed in a wide range of Italian language styles and contexts.
- **Gemma 2B Architecture**: Built on the powerful Gemma 2B framework, known for its efficiency and effectiveness in language generation tasks.
- **AutoTrain Enhanced**: Benefits from AutoTrain's optimization, making the model both robust and versatile in handling Italian text generation.

## Usage

Here's how to use this model for generating Italian text:

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model_path = "theoracle/gemma_italian_camoscio"
tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
    model_path,
    device_map="auto",
    torch_dtype='auto'
).eval()

# Example: Generating Italian text
prompt = "Inizia la storia con una giornata soleggiata in Sicilia, dove"

# Tokenize and generate text
encoding = tokenizer(prompt, return_tensors='pt', padding=True, truncation=True, max_length=500, add_special_tokens=True)
input_ids = encoding['input_ids']
attention_mask = encoding['attention_mask']

output_ids = model.generate(
    input_ids.to('cuda'),
    attention_mask=attention_mask.to('cuda'),
    max_new_tokens=300,
    pad_token_id=tokenizer.eos_token_id
)

generated_text = tokenizer.decode(output_ids[0], skip_special_tokens=True)
print(generated_text)
```

## Application Scenarios

This model is ideal for:
- Content creators looking to produce Italian-language articles, stories, or scripts.
- Developers creating conversational AI applications in Italian.
- Educators and language learners seeking tools for Italian language practice and improvement.

## Training and Technology

The `theoracle/gemma_italian_camoscio` model is trained using the AutoTrain platform for optimal performance, ensuring that it is well-suited for a broad spectrum of Italian text generation tasks. The Camoscio dataset provides a solid foundation, offering diverse and extensive coverage of the Italian language, which, combined with the Gemma 2B architecture, enables the model to generate coherent, nuanced, and contextually relevant Italian text.

## License

This model is available under an "other" license, facilitating its use in a wide array of applications. Users are encouraged to review the license terms to ensure compliance with their project requirements and intended use cases.