Training in progress, epoch 1
Browse files
    	
        adapter_config.json
    CHANGED
    
    | @@ -13,7 +13,7 @@ | |
| 13 | 
             
              "layers_pattern": null,
         | 
| 14 | 
             
              "layers_to_transform": null,
         | 
| 15 | 
             
              "loftq_config": {},
         | 
| 16 | 
            -
              "lora_alpha":  | 
| 17 | 
             
              "lora_bias": false,
         | 
| 18 | 
             
              "lora_dropout": 0.1,
         | 
| 19 | 
             
              "megatron_config": null,
         | 
| @@ -24,13 +24,13 @@ | |
| 24 | 
             
              "rank_pattern": {},
         | 
| 25 | 
             
              "revision": null,
         | 
| 26 | 
             
              "target_modules": [
         | 
| 27 | 
            -
                " | 
| 28 | 
            -
                "k_proj",
         | 
| 29 | 
            -
                "q_proj",
         | 
| 30 | 
            -
                "up_proj",
         | 
| 31 | 
             
                "down_proj",
         | 
|  | |
|  | |
| 32 | 
             
                "gate_proj",
         | 
| 33 | 
            -
                " | 
|  | |
| 34 | 
             
              ],
         | 
| 35 | 
             
              "task_type": "CAUSAL_LM",
         | 
| 36 | 
             
              "trainable_token_indices": null,
         | 
|  | |
| 13 | 
             
              "layers_pattern": null,
         | 
| 14 | 
             
              "layers_to_transform": null,
         | 
| 15 | 
             
              "loftq_config": {},
         | 
| 16 | 
            +
              "lora_alpha": 4,
         | 
| 17 | 
             
              "lora_bias": false,
         | 
| 18 | 
             
              "lora_dropout": 0.1,
         | 
| 19 | 
             
              "megatron_config": null,
         | 
|  | |
| 24 | 
             
              "rank_pattern": {},
         | 
| 25 | 
             
              "revision": null,
         | 
| 26 | 
             
              "target_modules": [
         | 
| 27 | 
            +
                "o_proj",
         | 
|  | |
|  | |
|  | |
| 28 | 
             
                "down_proj",
         | 
| 29 | 
            +
                "k_proj",
         | 
| 30 | 
            +
                "v_proj",
         | 
| 31 | 
             
                "gate_proj",
         | 
| 32 | 
            +
                "up_proj",
         | 
| 33 | 
            +
                "q_proj"
         | 
| 34 | 
             
              ],
         | 
| 35 | 
             
              "task_type": "CAUSAL_LM",
         | 
| 36 | 
             
              "trainable_token_indices": null,
         | 
    	
        adapter_model.safetensors
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
             
            size 13093200
         | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:8fd9b74c8b64954b8770f45ba78513a708fe4a0faf22b1bb9672db60025d345c
         | 
| 3 | 
             
            size 13093200
         | 
    	
        runs/May20_10-26-42_a7435d519f39/events.out.tfevents.1747736830.a7435d519f39.2140.0
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:7c39897051a5e69b95425c91730627c1e424b61a0ee9e518f220488771456884
         | 
| 3 | 
            +
            size 4184
         | 
    	
        tokenizer_config.json
    CHANGED
    
    | @@ -51335,7 +51335,7 @@ | |
| 51335 | 
             
                "image_token": "<image_soft_token>"
         | 
| 51336 | 
             
              },
         | 
| 51337 | 
             
              "image_token": "<image_soft_token>",
         | 
| 51338 | 
            -
              "model_max_length":  | 
| 51339 | 
             
              "pad_token": "<pad>",
         | 
| 51340 | 
             
              "processor_class": "Gemma3Processor",
         | 
| 51341 | 
             
              "sp_model_kwargs": null,
         | 
|  | |
| 51335 | 
             
                "image_token": "<image_soft_token>"
         | 
| 51336 | 
             
              },
         | 
| 51337 | 
             
              "image_token": "<image_soft_token>",
         | 
| 51338 | 
            +
              "model_max_length": 128,
         | 
| 51339 | 
             
              "pad_token": "<pad>",
         | 
| 51340 | 
             
              "processor_class": "Gemma3Processor",
         | 
| 51341 | 
             
              "sp_model_kwargs": null,
         | 
    	
        training_args.bin
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
            -
            size  | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:8e3056b15e9125de2ce03e8987f5a82f10171a1cec36f73cd81da4f2f0241a92
         | 
| 3 | 
            +
            size 5688
         | 
