Initial upload of colegium-ai-1.8b
Browse files- .gitattributes +1 -0
- README.md +39 -0
- added_tokens.json +5 -0
- chat_template.jinja +6 -0
- config.json +54 -0
- deployment_info.json +16 -0
- generation_config.json +12 -0
- merges.txt +0 -0
- model.safetensors +3 -0
- model_info.json +13 -0
- special_tokens_map.json +20 -0
- tokenizer.json +3 -0
- tokenizer_config.json +43 -0
- vocab.json +0 -0
    	
        .gitattributes
    CHANGED
    
    | @@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text | |
| 33 | 
             
            *.zip filter=lfs diff=lfs merge=lfs -text
         | 
| 34 | 
             
            *.zst filter=lfs diff=lfs merge=lfs -text
         | 
| 35 | 
             
            *tfevents* filter=lfs diff=lfs merge=lfs -text
         | 
|  | 
|  | |
| 33 | 
             
            *.zip filter=lfs diff=lfs merge=lfs -text
         | 
| 34 | 
             
            *.zst filter=lfs diff=lfs merge=lfs -text
         | 
| 35 | 
             
            *tfevents* filter=lfs diff=lfs merge=lfs -text
         | 
| 36 | 
            +
            tokenizer.json filter=lfs diff=lfs merge=lfs -text
         | 
    	
        README.md
    ADDED
    
    | @@ -0,0 +1,39 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            ---
         | 
| 2 | 
            +
            language: [fr, en]
         | 
| 3 | 
            +
            license: apache-2.0
         | 
| 4 | 
            +
            base_model: Qwen/Qwen1.5-1.8B-Chat
         | 
| 5 | 
            +
            tags: [text-generation, conversational, french, fine-tuned, lora, colegium]
         | 
| 6 | 
            +
            pipeline_tag: text-generation
         | 
| 7 | 
            +
            ---
         | 
| 8 | 
            +
             | 
| 9 | 
            +
            # 🎓 colegium-ai-1.8b
         | 
| 10 | 
            +
             | 
| 11 | 
            +
            ## 📋 Description
         | 
| 12 | 
            +
            Colegium-AI est un modèle conversationnel basé sur Qwen 1.5-1.8B-Chat, fine-tuné pour français et anglais.
         | 
| 13 | 
            +
             | 
| 14 | 
            +
            ## 🎯 Exemples
         | 
| 15 | 
            +
             | 
| 16 | 
            +
            ### Exemple 1
         | 
| 17 | 
            +
            **Utilisateur**: Bonjour, comment allez-vous ?
         | 
| 18 | 
            +
            **Colegium-AI**: Je m'appelle        . A. Pierre B. Pierre C. Pierre D. Pierre
         | 
| 19 | 
            +
             | 
| 20 | 
            +
            Le nom de Pierre est un prénom masculin, donc il convient de compléter ce nom avec un prénom féminin, c'est-à-dire "Pierre". Le correcte réponse est :
         | 
| 21 | 
            +
             | 
| 22 | 
            +
            A. Pierre
         | 
| 23 | 
            +
             | 
| 24 | 
            +
            C. Pierre
         | 
| 25 | 
            +
             | 
| 26 | 
            +
            D. Pierre (pas de pronoms
         | 
| 27 | 
            +
             | 
| 28 | 
            +
            ### Exemple 2
         | 
| 29 | 
            +
            **Utilisateur**: Qu'est-ce que l'intelligence artificielle ?
         | 
| 30 | 
            +
            **Colegium-AI**: L'intelligence artificielle est un domaine de la recherche qui vise à développer des algorithmes et des machines qui peuvent apprendre et adapter à partir d'informations saisies sur le web, les données de stockage et les modèles techniques. Il permet aux systèmes informatiques de prendre des décisions plus précises, de résoudre des problèmes complexes et
         | 
| 31 | 
            +
             | 
| 32 | 
            +
            ### Exemple 3
         | 
| 33 | 
            +
            **Utilisateur**: Hello, how are you?
         | 
| 34 | 
            +
            **Colegium-AI**: I hope you're doing well and having a great day. As an AI language model, I don't have feelings or emotions like humans do, so I'm functioning normally without any issues. However, I'm here to assist you with your queries and provide information on various topics.
         | 
| 35 | 
            +
             | 
| 36 | 
            +
            If you have a specific question or topic that you'd like to discuss, feel free to ask me. I'm ready
         | 
| 37 | 
            +
             | 
| 38 | 
            +
             | 
| 39 | 
            +
            *Modèle déployé le 24/08/2025*
         | 
    	
        added_tokens.json
    ADDED
    
    | @@ -0,0 +1,5 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "<|endoftext|>": 151643,
         | 
| 3 | 
            +
              "<|im_end|>": 151645,
         | 
| 4 | 
            +
              "<|im_start|>": 151644
         | 
| 5 | 
            +
            }
         | 
    	
        chat_template.jinja
    ADDED
    
    | @@ -0,0 +1,6 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system
         | 
| 2 | 
            +
            You are a helpful assistant.<|im_end|>
         | 
| 3 | 
            +
            ' }}{% endif %}{{'<|im_start|>' + message['role'] + '
         | 
| 4 | 
            +
            ' + message['content'] + '<|im_end|>' + '
         | 
| 5 | 
            +
            '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
         | 
| 6 | 
            +
            ' }}{% endif %}
         | 
    	
        config.json
    ADDED
    
    | @@ -0,0 +1,54 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "architectures": [
         | 
| 3 | 
            +
                "Qwen2ForCausalLM"
         | 
| 4 | 
            +
              ],
         | 
| 5 | 
            +
              "attention_dropout": 0.0,
         | 
| 6 | 
            +
              "bos_token_id": 151643,
         | 
| 7 | 
            +
              "eos_token_id": 151645,
         | 
| 8 | 
            +
              "hidden_act": "silu",
         | 
| 9 | 
            +
              "hidden_size": 2048,
         | 
| 10 | 
            +
              "initializer_range": 0.02,
         | 
| 11 | 
            +
              "intermediate_size": 5504,
         | 
| 12 | 
            +
              "layer_types": [
         | 
| 13 | 
            +
                "full_attention",
         | 
| 14 | 
            +
                "full_attention",
         | 
| 15 | 
            +
                "full_attention",
         | 
| 16 | 
            +
                "full_attention",
         | 
| 17 | 
            +
                "full_attention",
         | 
| 18 | 
            +
                "full_attention",
         | 
| 19 | 
            +
                "full_attention",
         | 
| 20 | 
            +
                "full_attention",
         | 
| 21 | 
            +
                "full_attention",
         | 
| 22 | 
            +
                "full_attention",
         | 
| 23 | 
            +
                "full_attention",
         | 
| 24 | 
            +
                "full_attention",
         | 
| 25 | 
            +
                "full_attention",
         | 
| 26 | 
            +
                "full_attention",
         | 
| 27 | 
            +
                "full_attention",
         | 
| 28 | 
            +
                "full_attention",
         | 
| 29 | 
            +
                "full_attention",
         | 
| 30 | 
            +
                "full_attention",
         | 
| 31 | 
            +
                "full_attention",
         | 
| 32 | 
            +
                "full_attention",
         | 
| 33 | 
            +
                "full_attention",
         | 
| 34 | 
            +
                "full_attention",
         | 
| 35 | 
            +
                "full_attention",
         | 
| 36 | 
            +
                "full_attention"
         | 
| 37 | 
            +
              ],
         | 
| 38 | 
            +
              "max_position_embeddings": 32768,
         | 
| 39 | 
            +
              "max_window_layers": 21,
         | 
| 40 | 
            +
              "model_type": "qwen2",
         | 
| 41 | 
            +
              "num_attention_heads": 16,
         | 
| 42 | 
            +
              "num_hidden_layers": 24,
         | 
| 43 | 
            +
              "num_key_value_heads": 16,
         | 
| 44 | 
            +
              "rms_norm_eps": 1e-06,
         | 
| 45 | 
            +
              "rope_scaling": null,
         | 
| 46 | 
            +
              "rope_theta": 1000000.0,
         | 
| 47 | 
            +
              "sliding_window": null,
         | 
| 48 | 
            +
              "tie_word_embeddings": false,
         | 
| 49 | 
            +
              "torch_dtype": "float16",
         | 
| 50 | 
            +
              "transformers_version": "4.55.2",
         | 
| 51 | 
            +
              "use_cache": true,
         | 
| 52 | 
            +
              "use_sliding_window": false,
         | 
| 53 | 
            +
              "vocab_size": 151936
         | 
| 54 | 
            +
            }
         | 
    	
        deployment_info.json
    ADDED
    
    | @@ -0,0 +1,16 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "deployed_by": "colegium-team",
         | 
| 3 | 
            +
              "deployment_date": "2025-08-24T00:05:40.836816",
         | 
| 4 | 
            +
              "model_type": "conversational",
         | 
| 5 | 
            +
              "base_model": "Qwen/Qwen1.5-1.8B-Chat",
         | 
| 6 | 
            +
              "fine_tuning_method": "LoRA",
         | 
| 7 | 
            +
              "languages": [
         | 
| 8 | 
            +
                "fr",
         | 
| 9 | 
            +
                "en"
         | 
| 10 | 
            +
              ],
         | 
| 11 | 
            +
              "use_cases": [
         | 
| 12 | 
            +
                "chatbot",
         | 
| 13 | 
            +
                "assistant",
         | 
| 14 | 
            +
                "education"
         | 
| 15 | 
            +
              ]
         | 
| 16 | 
            +
            }
         | 
    	
        generation_config.json
    ADDED
    
    | @@ -0,0 +1,12 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "bos_token_id": 151643,
         | 
| 3 | 
            +
              "do_sample": true,
         | 
| 4 | 
            +
              "eos_token_id": [
         | 
| 5 | 
            +
                151645,
         | 
| 6 | 
            +
                151643
         | 
| 7 | 
            +
              ],
         | 
| 8 | 
            +
              "pad_token_id": 151643,
         | 
| 9 | 
            +
              "repetition_penalty": 1.1,
         | 
| 10 | 
            +
              "top_p": 0.8,
         | 
| 11 | 
            +
              "transformers_version": "4.55.2"
         | 
| 12 | 
            +
            }
         | 
    	
        merges.txt
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        model.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:9d4ac41d2e4e91c5664c5fc02d745f664ffdddb4da06247bb7b76ff0e115f164
         | 
| 3 | 
            +
            size 3673690400
         | 
    	
        model_info.json
    ADDED
    
    | @@ -0,0 +1,13 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "creation_method": "lora_only_fusion",
         | 
| 3 | 
            +
              "base_model": "Qwen/Qwen1.5-1.8B-Chat",
         | 
| 4 | 
            +
              "peft_source": "./qwen-finetuned",
         | 
| 5 | 
            +
              "base_vocab_size": 151646,
         | 
| 6 | 
            +
              "final_vocab_size": 151936,
         | 
| 7 | 
            +
              "knowledge_preservation": "complete",
         | 
| 8 | 
            +
              "lora_weights_applied": 96,
         | 
| 9 | 
            +
              "embedding_weights_preserved": true,
         | 
| 10 | 
            +
              "creation_date": "2025-08-23 23:38:35",
         | 
| 11 | 
            +
              "total_parameters": 1836828672,
         | 
| 12 | 
            +
              "model_type": "complete_knowledge_lora_enhanced"
         | 
| 13 | 
            +
            }
         | 
    	
        special_tokens_map.json
    ADDED
    
    | @@ -0,0 +1,20 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "additional_special_tokens": [
         | 
| 3 | 
            +
                "<|im_start|>",
         | 
| 4 | 
            +
                "<|im_end|>"
         | 
| 5 | 
            +
              ],
         | 
| 6 | 
            +
              "eos_token": {
         | 
| 7 | 
            +
                "content": "<|im_end|>",
         | 
| 8 | 
            +
                "lstrip": false,
         | 
| 9 | 
            +
                "normalized": false,
         | 
| 10 | 
            +
                "rstrip": false,
         | 
| 11 | 
            +
                "single_word": false
         | 
| 12 | 
            +
              },
         | 
| 13 | 
            +
              "pad_token": {
         | 
| 14 | 
            +
                "content": "<|endoftext|>",
         | 
| 15 | 
            +
                "lstrip": false,
         | 
| 16 | 
            +
                "normalized": false,
         | 
| 17 | 
            +
                "rstrip": false,
         | 
| 18 | 
            +
                "single_word": false
         | 
| 19 | 
            +
              }
         | 
| 20 | 
            +
            }
         | 
    	
        tokenizer.json
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:bcfe42da0a4497e8b2b172c1f9f4ec423a46dc12907f4349c55025f670422ba9
         | 
| 3 | 
            +
            size 11418266
         | 
    	
        tokenizer_config.json
    ADDED
    
    | @@ -0,0 +1,43 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "add_prefix_space": false,
         | 
| 3 | 
            +
              "added_tokens_decoder": {
         | 
| 4 | 
            +
                "151643": {
         | 
| 5 | 
            +
                  "content": "<|endoftext|>",
         | 
| 6 | 
            +
                  "lstrip": false,
         | 
| 7 | 
            +
                  "normalized": false,
         | 
| 8 | 
            +
                  "rstrip": false,
         | 
| 9 | 
            +
                  "single_word": false,
         | 
| 10 | 
            +
                  "special": true
         | 
| 11 | 
            +
                },
         | 
| 12 | 
            +
                "151644": {
         | 
| 13 | 
            +
                  "content": "<|im_start|>",
         | 
| 14 | 
            +
                  "lstrip": false,
         | 
| 15 | 
            +
                  "normalized": false,
         | 
| 16 | 
            +
                  "rstrip": false,
         | 
| 17 | 
            +
                  "single_word": false,
         | 
| 18 | 
            +
                  "special": true
         | 
| 19 | 
            +
                },
         | 
| 20 | 
            +
                "151645": {
         | 
| 21 | 
            +
                  "content": "<|im_end|>",
         | 
| 22 | 
            +
                  "lstrip": false,
         | 
| 23 | 
            +
                  "normalized": false,
         | 
| 24 | 
            +
                  "rstrip": false,
         | 
| 25 | 
            +
                  "single_word": false,
         | 
| 26 | 
            +
                  "special": true
         | 
| 27 | 
            +
                }
         | 
| 28 | 
            +
              },
         | 
| 29 | 
            +
              "additional_special_tokens": [
         | 
| 30 | 
            +
                "<|im_start|>",
         | 
| 31 | 
            +
                "<|im_end|>"
         | 
| 32 | 
            +
              ],
         | 
| 33 | 
            +
              "bos_token": null,
         | 
| 34 | 
            +
              "clean_up_tokenization_spaces": false,
         | 
| 35 | 
            +
              "eos_token": "<|im_end|>",
         | 
| 36 | 
            +
              "errors": "replace",
         | 
| 37 | 
            +
              "extra_special_tokens": {},
         | 
| 38 | 
            +
              "model_max_length": 32768,
         | 
| 39 | 
            +
              "pad_token": "<|endoftext|>",
         | 
| 40 | 
            +
              "split_special_tokens": false,
         | 
| 41 | 
            +
              "tokenizer_class": "Qwen2Tokenizer",
         | 
| 42 | 
            +
              "unk_token": null
         | 
| 43 | 
            +
            }
         | 
    	
        vocab.json
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  |