asafaya commited on
Commit
c58dd10
1 Parent(s): a666c50

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -7,7 +7,7 @@ language:
7
  - tr
8
  pipeline_tag: text-generation
9
  widget:
10
- - text: "Benim adım Zeynep, ve en sevdiğim kitap"
11
  example_title: "Benim adım Zeynep"
12
  - text: "Bugünkü yemeğimiz"
13
  example_title: "Bugünkü yemeğimiz"
@@ -15,7 +15,7 @@ widget:
15
 
16
  # Kanarya-750M: Turkish Language Model
17
 
18
- <img src="https://asafaya.me/images/kanarya.webp" alt="Kanarya Logo" style="width:500px;"/>
19
 
20
  **Kanarya** is a pre-trained Turkish GPT-J 750M model. Released as part of [Turkish Data Depository](https://tdd.ai/) efforts, the Kanarya family has two versions (Kanarya-2B, Kanarya-0.7B). Kanarya-2B is the larger version and Kanarya-0.7B is the smaller version. Both models are trained on a large-scale Turkish text corpus, filtered from OSCAR and mC4 datasets. The training data is collected from various sources, including news, articles, and websites, to create a diverse and high-quality dataset. The models are trained using a JAX/Flax implementation of the [GPT-J](https://github.com/kingoflolz/mesh-transformer-jax) architecture. The models are only pre-trained and are intended to be fine-tuned on a wide range of Turkish NLP tasks.
21
 
 
7
  - tr
8
  pipeline_tag: text-generation
9
  widget:
10
+ - text: "Benim adım Zeynep, ve en sevdiğim kitabın adı:"
11
  example_title: "Benim adım Zeynep"
12
  - text: "Bugünkü yemeğimiz"
13
  example_title: "Bugünkü yemeğimiz"
 
15
 
16
  # Kanarya-750M: Turkish Language Model
17
 
18
+ <img src="https://asafaya.me/images/kanarya.webp" alt="Kanarya Logo" style="width:600px;"/>
19
 
20
  **Kanarya** is a pre-trained Turkish GPT-J 750M model. Released as part of [Turkish Data Depository](https://tdd.ai/) efforts, the Kanarya family has two versions (Kanarya-2B, Kanarya-0.7B). Kanarya-2B is the larger version and Kanarya-0.7B is the smaller version. Both models are trained on a large-scale Turkish text corpus, filtered from OSCAR and mC4 datasets. The training data is collected from various sources, including news, articles, and websites, to create a diverse and high-quality dataset. The models are trained using a JAX/Flax implementation of the [GPT-J](https://github.com/kingoflolz/mesh-transformer-jax) architecture. The models are only pre-trained and are intended to be fine-tuned on a wide range of Turkish NLP tasks.
21