Muyumba commited on
Commit
4b46763
·
verified ·
1 Parent(s): effcde2

Initial upload of colegium-ai-1.8b

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: [fr, en]
3
+ license: apache-2.0
4
+ base_model: Qwen/Qwen1.5-1.8B-Chat
5
+ tags: [text-generation, conversational, french, fine-tuned, lora, colegium]
6
+ pipeline_tag: text-generation
7
+ ---
8
+
9
+ # 🎓 colegium-ai-1.8b
10
+
11
+ ## 📋 Description
12
+ Colegium-AI est un modèle conversationnel basé sur Qwen 1.5-1.8B-Chat, fine-tuné pour français et anglais.
13
+
14
+ ## 🎯 Exemples
15
+
16
+ ### Exemple 1
17
+ **Utilisateur**: Bonjour, comment allez-vous ?
18
+ **Colegium-AI**: Je m'appelle . A. Pierre B. Pierre C. Pierre D. Pierre
19
+
20
+ Le nom de Pierre est un prénom masculin, donc il convient de compléter ce nom avec un prénom féminin, c'est-à-dire "Pierre". Le correcte réponse est :
21
+
22
+ A. Pierre
23
+
24
+ C. Pierre
25
+
26
+ D. Pierre (pas de pronoms
27
+
28
+ ### Exemple 2
29
+ **Utilisateur**: Qu'est-ce que l'intelligence artificielle ?
30
+ **Colegium-AI**: L'intelligence artificielle est un domaine de la recherche qui vise à développer des algorithmes et des machines qui peuvent apprendre et adapter à partir d'informations saisies sur le web, les données de stockage et les modèles techniques. Il permet aux systèmes informatiques de prendre des décisions plus précises, de résoudre des problèmes complexes et
31
+
32
+ ### Exemple 3
33
+ **Utilisateur**: Hello, how are you?
34
+ **Colegium-AI**: I hope you're doing well and having a great day. As an AI language model, I don't have feelings or emotions like humans do, so I'm functioning normally without any issues. However, I'm here to assist you with your queries and provide information on various topics.
35
+
36
+ If you have a specific question or topic that you'd like to discuss, feel free to ask me. I'm ready
37
+
38
+
39
+ *Modèle déployé le 24/08/2025*
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
chat_template.jinja ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system
2
+ You are a helpful assistant.<|im_end|>
3
+ ' }}{% endif %}{{'<|im_start|>' + message['role'] + '
4
+ ' + message['content'] + '<|im_end|>' + '
5
+ '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
6
+ ' }}{% endif %}
config.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 2048,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 5504,
12
+ "layer_types": [
13
+ "full_attention",
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention"
37
+ ],
38
+ "max_position_embeddings": 32768,
39
+ "max_window_layers": 21,
40
+ "model_type": "qwen2",
41
+ "num_attention_heads": 16,
42
+ "num_hidden_layers": 24,
43
+ "num_key_value_heads": 16,
44
+ "rms_norm_eps": 1e-06,
45
+ "rope_scaling": null,
46
+ "rope_theta": 1000000.0,
47
+ "sliding_window": null,
48
+ "tie_word_embeddings": false,
49
+ "torch_dtype": "float16",
50
+ "transformers_version": "4.55.2",
51
+ "use_cache": true,
52
+ "use_sliding_window": false,
53
+ "vocab_size": 151936
54
+ }
deployment_info.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "deployed_by": "colegium-team",
3
+ "deployment_date": "2025-08-24T00:05:40.836816",
4
+ "model_type": "conversational",
5
+ "base_model": "Qwen/Qwen1.5-1.8B-Chat",
6
+ "fine_tuning_method": "LoRA",
7
+ "languages": [
8
+ "fr",
9
+ "en"
10
+ ],
11
+ "use_cases": [
12
+ "chatbot",
13
+ "assistant",
14
+ "education"
15
+ ]
16
+ }
generation_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 151645,
6
+ 151643
7
+ ],
8
+ "pad_token_id": 151643,
9
+ "repetition_penalty": 1.1,
10
+ "top_p": 0.8,
11
+ "transformers_version": "4.55.2"
12
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d4ac41d2e4e91c5664c5fc02d745f664ffdddb4da06247bb7b76ff0e115f164
3
+ size 3673690400
model_info.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "creation_method": "lora_only_fusion",
3
+ "base_model": "Qwen/Qwen1.5-1.8B-Chat",
4
+ "peft_source": "./qwen-finetuned",
5
+ "base_vocab_size": 151646,
6
+ "final_vocab_size": 151936,
7
+ "knowledge_preservation": "complete",
8
+ "lora_weights_applied": 96,
9
+ "embedding_weights_preserved": true,
10
+ "creation_date": "2025-08-23 23:38:35",
11
+ "total_parameters": 1836828672,
12
+ "model_type": "complete_knowledge_lora_enhanced"
13
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "eos_token": {
7
+ "content": "<|im_end|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "pad_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ }
20
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
3
+ size 11418266
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "clean_up_tokenization_spaces": false,
35
+ "eos_token": "<|im_end|>",
36
+ "errors": "replace",
37
+ "extra_special_tokens": {},
38
+ "model_max_length": 32768,
39
+ "pad_token": "<|endoftext|>",
40
+ "split_special_tokens": false,
41
+ "tokenizer_class": "Qwen2Tokenizer",
42
+ "unk_token": null
43
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff