diff --git a/loras/code-switching-corrupted/en-de/adapter_config.json b/loras/code-switching-corrupted/en-de/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/code-switching-corrupted/en-de/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching-corrupted/en-de/head_config.json b/loras/code-switching-corrupted/en-de/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/code-switching-corrupted/en-de/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching-corrupted/en-de/pytorch_adapter.bin b/loras/code-switching-corrupted/en-de/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..0994d95440a616fada642859f8c151932238a357 --- /dev/null +++ b/loras/code-switching-corrupted/en-de/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0fe662031b8b9039cb2dbbf2ca70719c0ec08ec8c9c8961793b55f871ac544a9 +size 1333391 diff --git a/loras/code-switching-corrupted/en-de/pytorch_model_head.bin b/loras/code-switching-corrupted/en-de/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..59d4cff3ab6d172cedfea00ef40c7bdc26aac407 --- /dev/null +++ b/loras/code-switching-corrupted/en-de/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68cd7d4ba860567bc4fcacbce6ebb1a53efc6fb98fadea7cd98d5e22ee9e00c6 +size 342547 diff --git a/loras/code-switching-corrupted/es-en/adapter_config.json b/loras/code-switching-corrupted/es-en/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/code-switching-corrupted/es-en/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching-corrupted/es-en/head_config.json b/loras/code-switching-corrupted/es-en/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/code-switching-corrupted/es-en/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching-corrupted/es-en/pytorch_adapter.bin b/loras/code-switching-corrupted/es-en/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..2b0cb4f78584784f185e227cde8dc2d31c03e4f7 --- /dev/null +++ b/loras/code-switching-corrupted/es-en/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43547ffa2e6c290d0b07ca2eeffa36a392f9f4de91c1cae6718d3451740b1d6f +size 1333391 diff --git a/loras/code-switching-corrupted/es-en/pytorch_model_head.bin b/loras/code-switching-corrupted/es-en/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..c9dd0857facf4efecdafa2394f541f80f3eb2238 --- /dev/null +++ b/loras/code-switching-corrupted/es-en/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02fff4bb4a821eb02c82ad73c70beeda6851ef10d48b9782113aedab63462967 +size 342547 diff --git a/loras/code-switching-corrupted/tr-de/adapter_config.json b/loras/code-switching-corrupted/tr-de/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/code-switching-corrupted/tr-de/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching-corrupted/tr-de/head_config.json b/loras/code-switching-corrupted/tr-de/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/code-switching-corrupted/tr-de/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching-corrupted/tr-de/pytorch_adapter.bin b/loras/code-switching-corrupted/tr-de/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..63422c619d80362c77cf37eb78cb0c4d166befd6 --- /dev/null +++ b/loras/code-switching-corrupted/tr-de/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7b8c966505b7dd9cbc0ed6c2f9203873407c9fe6fd76324676d3a551bf66b82 +size 1333391 diff --git a/loras/code-switching-corrupted/tr-de/pytorch_model_head.bin b/loras/code-switching-corrupted/tr-de/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..f74583520095a4e279534de1f52bef89e13be47c --- /dev/null +++ b/loras/code-switching-corrupted/tr-de/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:069be2c44fef08e7bd2fd51a1e337e1aadabf0d8885002415e332e677db3220b +size 342547 diff --git a/loras/code-switching-corrupted/vi-en/adapter_config.json b/loras/code-switching-corrupted/vi-en/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/code-switching-corrupted/vi-en/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching-corrupted/vi-en/head_config.json b/loras/code-switching-corrupted/vi-en/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/code-switching-corrupted/vi-en/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching-corrupted/vi-en/pytorch_adapter.bin b/loras/code-switching-corrupted/vi-en/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..9645bd934524627c6ca193f29a1c19b92c38b207 --- /dev/null +++ b/loras/code-switching-corrupted/vi-en/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f458fd42a8eb8e63dd79453d311199263475a39bcd9d26a2258726c9f1222708 +size 1333391 diff --git a/loras/code-switching-corrupted/vi-en/pytorch_model_head.bin b/loras/code-switching-corrupted/vi-en/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..70680ba5a3f0efd0bb256b74088365cfa015b172 --- /dev/null +++ b/loras/code-switching-corrupted/vi-en/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04821442d193507ceef7c2a5b52c0104da3ed6c518c4cb10ccaef31ec0e006c1 +size 342547 diff --git a/loras/code-switching/en-de/adapter_config.json b/loras/code-switching/en-de/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/code-switching/en-de/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching/en-de/head_config.json b/loras/code-switching/en-de/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/code-switching/en-de/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching/en-de/pytorch_adapter.bin b/loras/code-switching/en-de/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..966116322e6b341f6023c99603a0269613e374e8 --- /dev/null +++ b/loras/code-switching/en-de/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69cd1616fffca223bbf4292b6151d44b1a36182e4e7063c06c60eec88639bcac +size 1333391 diff --git a/loras/code-switching/en-de/pytorch_model_head.bin b/loras/code-switching/en-de/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..b34404187306b840bb95328dba70386e1fe52199 --- /dev/null +++ b/loras/code-switching/en-de/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9cc1ec9ee633a553e9881026084beb073fba5fb626cff432963baa9f97cc9a2 +size 342547 diff --git a/loras/code-switching/es-en/adapter_config.json b/loras/code-switching/es-en/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/code-switching/es-en/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching/es-en/head_config.json b/loras/code-switching/es-en/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/code-switching/es-en/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching/es-en/pytorch_adapter.bin b/loras/code-switching/es-en/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..c4715e5505246fd9ce4fefb6ade40cf24eff6de9 --- /dev/null +++ b/loras/code-switching/es-en/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04d1899417f80cdbb79d33ed6f5fab7f7c1c911f0d7b5920c3e89a89bafeceae +size 1333391 diff --git a/loras/code-switching/es-en/pytorch_model_head.bin b/loras/code-switching/es-en/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8b8a80f45036da31c6782dbe2222d33915dd435c --- /dev/null +++ b/loras/code-switching/es-en/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff52ac00d05d0d6056722cfc7542986e2e95692fc44ed09986712ef51235ff76 +size 342547 diff --git a/loras/code-switching/tr-de/adapter_config.json b/loras/code-switching/tr-de/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/code-switching/tr-de/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching/tr-de/head_config.json b/loras/code-switching/tr-de/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/code-switching/tr-de/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching/tr-de/pytorch_adapter.bin b/loras/code-switching/tr-de/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..3c9c20e4e4d8a28d8125f67c2523e2ef1ea5bad1 --- /dev/null +++ b/loras/code-switching/tr-de/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ad7f8c38a393dc1296ddda62cca819f98626b4321c23c222be994e368aa5fdd +size 1333391 diff --git a/loras/code-switching/tr-de/pytorch_model_head.bin b/loras/code-switching/tr-de/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..a1580fe638cbd6fdead3ce09fec3a0f01aec9c17 --- /dev/null +++ b/loras/code-switching/tr-de/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1a411492ca39df981fbbad7b86da344d5c623abffcdd9f0c90e15935877e1e4 +size 342547 diff --git a/loras/code-switching/vi-en/adapter_config.json b/loras/code-switching/vi-en/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/code-switching/vi-en/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching/vi-en/head_config.json b/loras/code-switching/vi-en/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/code-switching/vi-en/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/code-switching/vi-en/pytorch_adapter.bin b/loras/code-switching/vi-en/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ff1f3ca3dc60180d2240c8de2f8c3ea4cc9d72b2 --- /dev/null +++ b/loras/code-switching/vi-en/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08b3a1385f0fccaf23ce071a2b1baa40f8b43e22bc5ebc3e8b4cc59fc4a2bbc0 +size 1333391 diff --git a/loras/code-switching/vi-en/pytorch_model_head.bin b/loras/code-switching/vi-en/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..bf365ec8d16e46166d07eca1645934d4c009ce52 --- /dev/null +++ b/loras/code-switching/vi-en/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5042d77c9ff52af2bd05b8cdb4ddbed6d667d90e4a445862ff54d549d748f9da +size 342547 diff --git a/loras/ersatz/ar/adapter_config.json b/loras/ersatz/ar/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/ar/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/ar/head_config.json b/loras/ersatz/ar/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/ar/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/ar/pytorch_adapter.bin b/loras/ersatz/ar/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ee7170f86222b32a97514bebc45865a3ea929b67 --- /dev/null +++ b/loras/ersatz/ar/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bface62e88d0620d993304491bc838fe4145dd124b2f7c6c3b0447c43a89d3bb +size 1333391 diff --git a/loras/ersatz/ar/pytorch_model_head.bin b/loras/ersatz/ar/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..64d09cf9f04e20627e33ad984e1c8f0134959f03 --- /dev/null +++ b/loras/ersatz/ar/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92e49102f467213dc01178eef57ae4ac2c827af96a5443060dff76267ea0dc16 +size 342547 diff --git a/loras/ersatz/cs/adapter_config.json b/loras/ersatz/cs/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/cs/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/cs/head_config.json b/loras/ersatz/cs/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/cs/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/cs/pytorch_adapter.bin b/loras/ersatz/cs/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..5bd0b175c4358cbad1af5f8ef4e2cfa089186552 --- /dev/null +++ b/loras/ersatz/cs/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a23d3d31fdbe1e2d50ff03b4517baf5c38d4069b9ce72d2df3567592bfbc873 +size 1333391 diff --git a/loras/ersatz/cs/pytorch_model_head.bin b/loras/ersatz/cs/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..0142ef1ab60d7dfb647b9f9bc5a31721a67fd0a6 --- /dev/null +++ b/loras/ersatz/cs/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c92efa387a97be592665ad6330ae05ef567c5ab63ed8b24b44d6b29d5351a21c +size 342547 diff --git a/loras/ersatz/de/adapter_config.json b/loras/ersatz/de/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/de/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/de/head_config.json b/loras/ersatz/de/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/de/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/de/pytorch_adapter.bin b/loras/ersatz/de/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..69793f1fe6dbb2aa0ce47e1f7b91a1f74b552820 --- /dev/null +++ b/loras/ersatz/de/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9078e3fe01f286a60b0cccb2491147ab417a8ca00fa18e94c492ee3098d0f82d +size 1333391 diff --git a/loras/ersatz/de/pytorch_model_head.bin b/loras/ersatz/de/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..63ac61444b3b65f3c04f908946b3c65b9fd8275f --- /dev/null +++ b/loras/ersatz/de/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:587a7b064bfb6b35fbc843fb9ec1eceb57dfd0534f81d9130b4aa0dfb6d3dd41 +size 342547 diff --git a/loras/ersatz/en/adapter_config.json b/loras/ersatz/en/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/en/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/en/head_config.json b/loras/ersatz/en/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/en/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/en/pytorch_adapter.bin b/loras/ersatz/en/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..945c96ae3222c148350194c731d5bab97764937f --- /dev/null +++ b/loras/ersatz/en/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81ee9fcf686e2ac3d0a9e85f67196da942743ec14c906e76db2dc4d0c9c6b391 +size 1333391 diff --git a/loras/ersatz/en/pytorch_model_head.bin b/loras/ersatz/en/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..06d93213b3a083ad635a9d4f6a646591dc781c1e --- /dev/null +++ b/loras/ersatz/en/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3ff56382de61fe4d1a6fcf2e71549420c6d0a9a56367f29a48748f01e02f8c9 +size 342547 diff --git a/loras/ersatz/et/adapter_config.json b/loras/ersatz/et/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/et/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/et/head_config.json b/loras/ersatz/et/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/et/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/et/pytorch_adapter.bin b/loras/ersatz/et/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..906d105fc753ea62ac19620c3ee76873d307b4c5 --- /dev/null +++ b/loras/ersatz/et/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:931f08a8a2dd34d1b938c98ab79cdc81775910857beeab9239163e2e743a77b7 +size 1333391 diff --git a/loras/ersatz/et/pytorch_model_head.bin b/loras/ersatz/et/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..e730a51ea4e966a4c7d49fd0068945ac08ca2916 --- /dev/null +++ b/loras/ersatz/et/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ba0c4c4aa7c9b06197db1afbb7663e24ee3b68e8ed59270ee31f725dc00588b +size 342547 diff --git a/loras/ersatz/fi/adapter_config.json b/loras/ersatz/fi/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/fi/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/fi/head_config.json b/loras/ersatz/fi/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/fi/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/fi/pytorch_adapter.bin b/loras/ersatz/fi/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..94549f62b5ef8fe1f6497244e765d9d0a9744090 --- /dev/null +++ b/loras/ersatz/fi/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:174e859c262f8d1e643699d89c18c8f9419a01abfb2e306a667a077334032198 +size 1333391 diff --git a/loras/ersatz/fi/pytorch_model_head.bin b/loras/ersatz/fi/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..dcbba902ffd80843ba5b3317fba9cb5b3e4f46ef --- /dev/null +++ b/loras/ersatz/fi/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:919bb8ef2340be93a1ba7a5e942bc9f31ad104734bb11cd70d5f585babdfa288 +size 342547 diff --git a/loras/ersatz/fr/adapter_config.json b/loras/ersatz/fr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/fr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/fr/head_config.json b/loras/ersatz/fr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/fr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/fr/pytorch_adapter.bin b/loras/ersatz/fr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d7e33408820333bf9aa7475db80e98c06cd7adef --- /dev/null +++ b/loras/ersatz/fr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:029b3190336ae6d0b82e54f2dd25b25603680da750c0a8cc3a1a7f2763a5c750 +size 1333391 diff --git a/loras/ersatz/fr/pytorch_model_head.bin b/loras/ersatz/fr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..28167c0390f7246f6ad699eaf84bbb1e10ae23c8 --- /dev/null +++ b/loras/ersatz/fr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2077fbe75748a709a0802313dafedc6ebe4ce58d0753d745a616b1775f5394c6 +size 342547 diff --git a/loras/ersatz/gu/adapter_config.json b/loras/ersatz/gu/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/gu/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/gu/head_config.json b/loras/ersatz/gu/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/gu/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/gu/pytorch_adapter.bin b/loras/ersatz/gu/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ea15d2edca7e970fb461f33c3cd3d83106347da7 --- /dev/null +++ b/loras/ersatz/gu/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1505d14f3919abe34f264fe4a30671b769a2be20ca59633f20e9a9a56489f2dc +size 1333391 diff --git a/loras/ersatz/gu/pytorch_model_head.bin b/loras/ersatz/gu/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..767ac1e05ba59272e6b44072ca3ac118a0a15b63 --- /dev/null +++ b/loras/ersatz/gu/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c468417b40e1451e2fd6c63774a25752abb1968b7bc99c0c529cb9092bbb170 +size 342547 diff --git a/loras/ersatz/hi/adapter_config.json b/loras/ersatz/hi/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/hi/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/hi/head_config.json b/loras/ersatz/hi/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/hi/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/hi/pytorch_adapter.bin b/loras/ersatz/hi/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..a841187a7ed34eb731ee0280b256004fb003b080 --- /dev/null +++ b/loras/ersatz/hi/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44bb5b47889fb16aa8b6d853a42d0a03eaad7a8132b9c40b48523ec9117435f3 +size 1333391 diff --git a/loras/ersatz/hi/pytorch_model_head.bin b/loras/ersatz/hi/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..1b3d282c4ad1a2780835d3e041395591ac036cc8 --- /dev/null +++ b/loras/ersatz/hi/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc2a47e620437f260320cc1673a1745d5b2fe84d955cd6161302ec6b28d689c6 +size 342547 diff --git a/loras/ersatz/ja/adapter_config.json b/loras/ersatz/ja/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/ja/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/ja/head_config.json b/loras/ersatz/ja/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/ja/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/ja/pytorch_adapter.bin b/loras/ersatz/ja/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..f35ffab447b924d861453d21763dc7af37345746 --- /dev/null +++ b/loras/ersatz/ja/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9a986a652862750853e369cc6dcff3427fdd5c5d3c7bd414e24e43711a2db52 +size 1333391 diff --git a/loras/ersatz/ja/pytorch_model_head.bin b/loras/ersatz/ja/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..cb71f5163fe7bd22643bcf08aa934e6ee64a0011 --- /dev/null +++ b/loras/ersatz/ja/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ad77bdf4ee6b287f8efa882b09e2221ec662b4c5d1d18e4788d1d11c6ec7169 +size 342547 diff --git a/loras/ersatz/kk/adapter_config.json b/loras/ersatz/kk/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/kk/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/kk/head_config.json b/loras/ersatz/kk/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/kk/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/kk/pytorch_adapter.bin b/loras/ersatz/kk/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..a16e6705043f1e4b3674e66d347db8f707c563e6 --- /dev/null +++ b/loras/ersatz/kk/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84bc36e6fb4df35daeb3c95768d3b3766a6eca453d2c0111dca0c6daa3c5e1ff +size 1333391 diff --git a/loras/ersatz/kk/pytorch_model_head.bin b/loras/ersatz/kk/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..f2a3118fc7856fa6d05fef0d96571b3ae1cf2567 --- /dev/null +++ b/loras/ersatz/kk/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73dcfcec54250396553926d6272bd01c2a6b1f7f06f38d2e3473efb8d3e52bc9 +size 342547 diff --git a/loras/ersatz/km/adapter_config.json b/loras/ersatz/km/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/km/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/km/head_config.json b/loras/ersatz/km/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/km/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/km/pytorch_adapter.bin b/loras/ersatz/km/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..10bb6598a279f3443bc3beb7b2503c938a1553f9 --- /dev/null +++ b/loras/ersatz/km/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42bc430d9b66ddc6bc5710de8843d2f84f5dafaee7c5aa8dfdfe00854663dbf8 +size 1333391 diff --git a/loras/ersatz/km/pytorch_model_head.bin b/loras/ersatz/km/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..d129ee3f22eda68b86434a96663778776b947a9b --- /dev/null +++ b/loras/ersatz/km/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:af617ef380db20472644447db6098f279fc7179ee70f2d92703b7a476db91d9e +size 342547 diff --git a/loras/ersatz/lt/adapter_config.json b/loras/ersatz/lt/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/lt/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/lt/head_config.json b/loras/ersatz/lt/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/lt/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/lt/pytorch_adapter.bin b/loras/ersatz/lt/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..e0f63039b8b40ede1529a70dcef76321948058d2 --- /dev/null +++ b/loras/ersatz/lt/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6bc698763c5d98eb94e8d35806efc12caba1c49963005adaefd551781c00b777 +size 1333391 diff --git a/loras/ersatz/lt/pytorch_model_head.bin b/loras/ersatz/lt/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..112ddb3d56bfdb44bc68f05c1658b40e1f02dce3 --- /dev/null +++ b/loras/ersatz/lt/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b65276adf3078e34be01dbc1106984a838fb95752552e5ecd784c5429d91f373 +size 342547 diff --git a/loras/ersatz/lv/adapter_config.json b/loras/ersatz/lv/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/lv/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/lv/head_config.json b/loras/ersatz/lv/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/lv/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/lv/pytorch_adapter.bin b/loras/ersatz/lv/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..cb0df87a3db9b3be1745112b142cbe3123b62f12 --- /dev/null +++ b/loras/ersatz/lv/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd1539a5f76b3c22772a3b3dfe421c680c2906f5b1d9e1198cb5c3bf770cb188 +size 1333391 diff --git a/loras/ersatz/lv/pytorch_model_head.bin b/loras/ersatz/lv/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..ed5f8fea174426702b5f6516dcfc776f5e6453ad --- /dev/null +++ b/loras/ersatz/lv/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e4a1abe69f15f6cabc3010cd34ea95df92f7bbb6d53ee23b5fbe2ca636f4327 +size 342547 diff --git a/loras/ersatz/pl/adapter_config.json b/loras/ersatz/pl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/pl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/pl/head_config.json b/loras/ersatz/pl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/pl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/pl/pytorch_adapter.bin b/loras/ersatz/pl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..895a54c26f3de4c379b2a4611c9b84e02e5d9b9a --- /dev/null +++ b/loras/ersatz/pl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6019ec683dc730b947a3f289170dc97293397eb53431defe559010f07efebe8e +size 1333391 diff --git a/loras/ersatz/pl/pytorch_model_head.bin b/loras/ersatz/pl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..6e0d30f40dd77b16023b217d483a469a341890ed --- /dev/null +++ b/loras/ersatz/pl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:166f65e835ee29f02a7c4cb2b6a26bdc8126760f5554ee510ac727ffb6b8527f +size 342547 diff --git a/loras/ersatz/ps/adapter_config.json b/loras/ersatz/ps/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/ps/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/ps/head_config.json b/loras/ersatz/ps/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/ps/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/ps/pytorch_adapter.bin b/loras/ersatz/ps/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..75a92177bae6afeb85db0c526bb69f5ced199d9c --- /dev/null +++ b/loras/ersatz/ps/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0de7120f399e95f82232f403779ae9a06415767dffe7dca24bde28fbf794999 +size 1333391 diff --git a/loras/ersatz/ps/pytorch_model_head.bin b/loras/ersatz/ps/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..9d50f629b2aea5fbe06cede450dcb7b99203d3f9 --- /dev/null +++ b/loras/ersatz/ps/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf66dd82d0d66edd37fec1fd6a8821b711f63eb156e86d6273a65c640ce2efee +size 342547 diff --git a/loras/ersatz/ro/adapter_config.json b/loras/ersatz/ro/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/ro/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/ro/head_config.json b/loras/ersatz/ro/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/ro/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/ro/pytorch_adapter.bin b/loras/ersatz/ro/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..61c039e325996f29182f8c7505715c9a28a0efbd --- /dev/null +++ b/loras/ersatz/ro/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf5431313308c1fded433cefdff08feedccb93ad6b8ff02981b87316efe1723f +size 1333391 diff --git a/loras/ersatz/ro/pytorch_model_head.bin b/loras/ersatz/ro/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..a833fe2256c3f95c7b1b3157f1c5df1406cf62f4 --- /dev/null +++ b/loras/ersatz/ro/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:003e5b3ea83f7a7cf9e812030e25e7d4cfeba10946b884094669a3bbfd3285b3 +size 342547 diff --git a/loras/ersatz/ru/adapter_config.json b/loras/ersatz/ru/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/ru/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/ru/head_config.json b/loras/ersatz/ru/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/ru/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/ru/pytorch_adapter.bin b/loras/ersatz/ru/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..3260aee9b74c80e2772d6c42bf43f096dcff3a3a --- /dev/null +++ b/loras/ersatz/ru/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3502f24e1d8e795a8e7e38c00d735d2edff184d46a05795644c412d844cc9d3d +size 1333391 diff --git a/loras/ersatz/ru/pytorch_model_head.bin b/loras/ersatz/ru/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..95d42625c42d507880330f3b4fc9e00d55cc4ab3 --- /dev/null +++ b/loras/ersatz/ru/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b86fc648a3148688ba932663d074824f9d3a945d23f43e53e96d0d7849a99bc +size 342547 diff --git a/loras/ersatz/ta/adapter_config.json b/loras/ersatz/ta/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/ta/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/ta/head_config.json b/loras/ersatz/ta/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/ta/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/ta/pytorch_adapter.bin b/loras/ersatz/ta/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..6318c9eb549796cc869f672811e76bdd558646e1 --- /dev/null +++ b/loras/ersatz/ta/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b12683185e2ea2134947627c74c6a611d0f91877958f1587776ac8f51851d52b +size 1333391 diff --git a/loras/ersatz/ta/pytorch_model_head.bin b/loras/ersatz/ta/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..0734834eea398afea1f123186c9f8d45ddf61824 --- /dev/null +++ b/loras/ersatz/ta/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e1cd009f9fa6df5df4244407670a90ee3acb023019be360c614d4c41ad97a4d +size 342547 diff --git a/loras/ersatz/tr/adapter_config.json b/loras/ersatz/tr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/tr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/tr/head_config.json b/loras/ersatz/tr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/tr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/tr/pytorch_adapter.bin b/loras/ersatz/tr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..2c14da1dbee182ebf8a2afe52a182ca630026c76 --- /dev/null +++ b/loras/ersatz/tr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b4c99ed4ad345ae48b4192ab8e59faab08e6e8b6276d301b6ff774af1abd3da +size 1333391 diff --git a/loras/ersatz/tr/pytorch_model_head.bin b/loras/ersatz/tr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..cd7295f8570fe4256ee18cf2cad02be77244dbb2 --- /dev/null +++ b/loras/ersatz/tr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42f61ed9c2aa71d7f25f0f8757327bde3b1b4b4f2506453fbee40d81526adb19 +size 342547 diff --git a/loras/ersatz/zh/adapter_config.json b/loras/ersatz/zh/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ersatz/zh/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/zh/head_config.json b/loras/ersatz/zh/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ersatz/zh/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ersatz/zh/pytorch_adapter.bin b/loras/ersatz/zh/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..e41d004175be625d28972887c6d70ce0aca9e50d --- /dev/null +++ b/loras/ersatz/zh/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae0a427c8fce2d0cee85d2f0ac726e5ff7012a5e385e38ab69c0074752083162 +size 1333391 diff --git a/loras/ersatz/zh/pytorch_model_head.bin b/loras/ersatz/zh/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..579b64dec0682a4699745cccc6f305c52b62bef4 --- /dev/null +++ b/loras/ersatz/zh/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59320f21f5072839d477c53ae8d26be2d6c609b5936569824270120cb2cd3be0 +size 342547 diff --git a/loras/legal-laws-corrupted/de/adapter_config.json b/loras/legal-laws-corrupted/de/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/legal-laws-corrupted/de/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws-corrupted/de/head_config.json b/loras/legal-laws-corrupted/de/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/legal-laws-corrupted/de/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws-corrupted/de/pytorch_adapter.bin b/loras/legal-laws-corrupted/de/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d30e3f5092e0d09ed197e5433b6a15feaff9e02a --- /dev/null +++ b/loras/legal-laws-corrupted/de/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:567f94f15afa245df2c65ea00ab884b0c8e86e0f34b052d9d06b0e7b89df0f1a +size 1333391 diff --git a/loras/legal-laws-corrupted/de/pytorch_model_head.bin b/loras/legal-laws-corrupted/de/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..087c049e227d3c08a2d22ec8991c36342c0d892f --- /dev/null +++ b/loras/legal-laws-corrupted/de/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c1f0e09a0d96f77ef6bb7bfac4b68e4f773bd6b89ccac0bcd53d6d2599a1b24 +size 342547 diff --git a/loras/legal-laws-corrupted/es/adapter_config.json b/loras/legal-laws-corrupted/es/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/legal-laws-corrupted/es/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws-corrupted/es/head_config.json b/loras/legal-laws-corrupted/es/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/legal-laws-corrupted/es/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws-corrupted/es/pytorch_adapter.bin b/loras/legal-laws-corrupted/es/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..019bf83eaed460c5f749b997ca5d26e58b806e3c --- /dev/null +++ b/loras/legal-laws-corrupted/es/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d063b559e768be57a0cfe09e264e2a90ba983785b80d8532b9c3345b8349990 +size 1333391 diff --git a/loras/legal-laws-corrupted/es/pytorch_model_head.bin b/loras/legal-laws-corrupted/es/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8144c5b314b76b5830a2f50b2296800e1ca8804f --- /dev/null +++ b/loras/legal-laws-corrupted/es/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85179b7fb01395bb5d732a5942b94f3346c16eb840b2d03f5bb77a9b953aa17f +size 342547 diff --git a/loras/legal-laws-corrupted/fr/adapter_config.json b/loras/legal-laws-corrupted/fr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/legal-laws-corrupted/fr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws-corrupted/fr/head_config.json b/loras/legal-laws-corrupted/fr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/legal-laws-corrupted/fr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws-corrupted/fr/pytorch_adapter.bin b/loras/legal-laws-corrupted/fr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..6c3837350bf902b6ea67a085073052f415d8895e --- /dev/null +++ b/loras/legal-laws-corrupted/fr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b278ea23d349fadfdadad6491114943be0fff2d7c26ebe57f7c521efe30c689b +size 1333391 diff --git a/loras/legal-laws-corrupted/fr/pytorch_model_head.bin b/loras/legal-laws-corrupted/fr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..e15d00a32b55f387f470037dcb1a7815eec14dba --- /dev/null +++ b/loras/legal-laws-corrupted/fr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:610c41eb2a8e949b8a92a661a197ebd45d041cd05be6909da0b60bf35325992f +size 342547 diff --git a/loras/legal-laws-corrupted/it/adapter_config.json b/loras/legal-laws-corrupted/it/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/legal-laws-corrupted/it/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws-corrupted/it/head_config.json b/loras/legal-laws-corrupted/it/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/legal-laws-corrupted/it/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws-corrupted/it/pytorch_adapter.bin b/loras/legal-laws-corrupted/it/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..eb092a573ec589fd85915eaf7fd9f0c348449ddb --- /dev/null +++ b/loras/legal-laws-corrupted/it/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b7f8e12d8694e22edcdcbb2a26d828de88c9061ed7c8df72683cce52cc6b01a +size 1333391 diff --git a/loras/legal-laws-corrupted/it/pytorch_model_head.bin b/loras/legal-laws-corrupted/it/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..7e90404ae37cdca28f0f8babdb5ab374b41829f4 --- /dev/null +++ b/loras/legal-laws-corrupted/it/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2afdcc5ef73bde0eef6c18ece1dd062f8e3103d5c846b69c857fe59998159bea +size 342547 diff --git a/loras/legal-laws-corrupted/pt/adapter_config.json b/loras/legal-laws-corrupted/pt/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/legal-laws-corrupted/pt/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws-corrupted/pt/head_config.json b/loras/legal-laws-corrupted/pt/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/legal-laws-corrupted/pt/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws-corrupted/pt/pytorch_adapter.bin b/loras/legal-laws-corrupted/pt/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1a2d3d9043c21874ed8bb284f30daee87d796917 --- /dev/null +++ b/loras/legal-laws-corrupted/pt/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd1c7eed73a7c45786ae522999a391c3354f68b14a4b3674315bad58be311aaa +size 1333391 diff --git a/loras/legal-laws-corrupted/pt/pytorch_model_head.bin b/loras/legal-laws-corrupted/pt/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..1899e6bcde61c42f3e78a12b43fa13fd1400dd0f --- /dev/null +++ b/loras/legal-laws-corrupted/pt/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46788e504e65587ed09d83aaee2aa7fb605ee52ba398744798c2a4601380d18d +size 342547 diff --git a/loras/legal-laws/de/adapter_config.json b/loras/legal-laws/de/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/legal-laws/de/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws/de/head_config.json b/loras/legal-laws/de/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/legal-laws/de/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws/de/pytorch_adapter.bin b/loras/legal-laws/de/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..2a8ce996c7b369c3286d1550e9b6297e6082c84d --- /dev/null +++ b/loras/legal-laws/de/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:268074fe418bf9613287dd8a4a5b0ae8c510a114638c10740d24bc3a7c03e9e0 +size 1333391 diff --git a/loras/legal-laws/de/pytorch_model_head.bin b/loras/legal-laws/de/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..3cf7345fd44f3bfd042f9433a2c650fb8a85ab30 --- /dev/null +++ b/loras/legal-laws/de/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:164ba380f5b6a06b61571aae19120b29a25e4a9b95216928f136948c1d0ce425 +size 342547 diff --git a/loras/legal-laws/es/adapter_config.json b/loras/legal-laws/es/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/legal-laws/es/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws/es/head_config.json b/loras/legal-laws/es/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/legal-laws/es/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws/es/pytorch_adapter.bin b/loras/legal-laws/es/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..0c46a9cb1cdab4648103a887c9f3989e1290a539 --- /dev/null +++ b/loras/legal-laws/es/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e152a8aac74441fecaf485b5a602e42b2d1ab296df91e4e1897a583abf0c7357 +size 1333391 diff --git a/loras/legal-laws/es/pytorch_model_head.bin b/loras/legal-laws/es/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..b19b810526f651a3a776f4f25f3ebd2c5f23e039 --- /dev/null +++ b/loras/legal-laws/es/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1169b90ec5eb83808607e5fa831309ac73984d17d23c54c300658506ffbe805e +size 342547 diff --git a/loras/legal-laws/fr/adapter_config.json b/loras/legal-laws/fr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/legal-laws/fr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws/fr/head_config.json b/loras/legal-laws/fr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/legal-laws/fr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws/fr/pytorch_adapter.bin b/loras/legal-laws/fr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1061ee424fa2571d9ff3b0f81ee38dc4731843f2 --- /dev/null +++ b/loras/legal-laws/fr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19191d52f79d6688d3e460663284e7b7b6df64f9259fa393e1212cb1a5462f62 +size 1333391 diff --git a/loras/legal-laws/fr/pytorch_model_head.bin b/loras/legal-laws/fr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..62b8824077716a548ad9fea07226646e2f9b9fe9 --- /dev/null +++ b/loras/legal-laws/fr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05a0e14a35e3e01e662c7d754f961df6f0c2e8643ddb648dcbecb17370bb65b7 +size 342547 diff --git a/loras/legal-laws/it/adapter_config.json b/loras/legal-laws/it/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/legal-laws/it/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws/it/head_config.json b/loras/legal-laws/it/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/legal-laws/it/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws/it/pytorch_adapter.bin b/loras/legal-laws/it/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..8495376d0672ee4458f8e771cbdfe7f80a4a233f --- /dev/null +++ b/loras/legal-laws/it/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c1bc74de056d6e53d7bf814e64f6826c7e2ef6d937e9a80a02fb09b1440a049 +size 1333391 diff --git a/loras/legal-laws/it/pytorch_model_head.bin b/loras/legal-laws/it/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..5a3de14b3efa0cc3ede598154c23962eb36c4054 --- /dev/null +++ b/loras/legal-laws/it/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91e1189491e2029e8faddf3723060b48c7c724f94cd74f7a305fe5098428c8d1 +size 342547 diff --git a/loras/legal-laws/pt/adapter_config.json b/loras/legal-laws/pt/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/legal-laws/pt/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws/pt/head_config.json b/loras/legal-laws/pt/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/legal-laws/pt/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/legal-laws/pt/pytorch_adapter.bin b/loras/legal-laws/pt/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..93a036bf6a2b0367c5e282c0a71a6e91aa03b59f --- /dev/null +++ b/loras/legal-laws/pt/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4f9623f2d97f37d22c4c2fab4807010c733f59d8427720f25b3ecd52ff32c28 +size 1333391 diff --git a/loras/legal-laws/pt/pytorch_model_head.bin b/loras/legal-laws/pt/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..c59286803e369ebfe5a61a1a9fbb5a47dcb3a566 --- /dev/null +++ b/loras/legal-laws/pt/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:16362cfcf042f651053949477d62cd383e09f2968cf5956dc1c2ab575bc23434 +size 342547 diff --git a/loras/opus100/af/adapter_config.json b/loras/opus100/af/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/af/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/af/head_config.json b/loras/opus100/af/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/af/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/af/pytorch_adapter.bin b/loras/opus100/af/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..7916b8541c58e67daccfc7ca256c092fed916467 --- /dev/null +++ b/loras/opus100/af/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93640fe53271eb3c2547ab515664461c97dbb356a239cc6472a4f4d635ca1bc2 +size 1333391 diff --git a/loras/opus100/af/pytorch_model_head.bin b/loras/opus100/af/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..232fabe13b492bc3e94000838d83eb271dcae533 --- /dev/null +++ b/loras/opus100/af/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0cdb6665814f84e45e243675df3419a64fc38d047f3458799e6220016c32cd2a +size 342547 diff --git a/loras/opus100/am/adapter_config.json b/loras/opus100/am/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/am/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/am/head_config.json b/loras/opus100/am/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/am/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/am/pytorch_adapter.bin b/loras/opus100/am/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..fdb9f4f585d930de38815a122b1345b12bec5ad6 --- /dev/null +++ b/loras/opus100/am/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e97ec40ffbc63279fb763bcd25966b55dc858a0c250bea693fdabcc8ce1daf6f +size 1333391 diff --git a/loras/opus100/am/pytorch_model_head.bin b/loras/opus100/am/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..1729f6bba8fbb61ede6ca4da0491f49d878f982f --- /dev/null +++ b/loras/opus100/am/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:039a05d804503628a373b303ea5ca4e10d600d2142471f2491b89a5fdc7cf26b +size 342547 diff --git a/loras/opus100/ar/adapter_config.json b/loras/opus100/ar/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ar/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ar/head_config.json b/loras/opus100/ar/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ar/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ar/pytorch_adapter.bin b/loras/opus100/ar/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..c123aa6190ba6f2a2868a47d3479531db369247d --- /dev/null +++ b/loras/opus100/ar/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5769713b5cfdc41eb91b522684e6e4c78b2f67be2f5b0e29285da9092f07dac6 +size 1333391 diff --git a/loras/opus100/ar/pytorch_model_head.bin b/loras/opus100/ar/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..7029fbf8e9394598ae8fba5b93e10a3c994a5fbe --- /dev/null +++ b/loras/opus100/ar/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9da9c54528b5086e4ac716a6900fd4059bb6d14f3f64ab2af7f77aabb480bf2d +size 342547 diff --git a/loras/opus100/az/adapter_config.json b/loras/opus100/az/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/az/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/az/head_config.json b/loras/opus100/az/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/az/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/az/pytorch_adapter.bin b/loras/opus100/az/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1675acb8f2ec168f50de822867a2b816aafdd01a --- /dev/null +++ b/loras/opus100/az/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1b715d8a6e24f8a192b7e200e49e567cfc635872ca5a2691991a078a69af02b +size 1333391 diff --git a/loras/opus100/az/pytorch_model_head.bin b/loras/opus100/az/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..66e9c4b94b19ae6e8989cf5986f426c48b4c0107 --- /dev/null +++ b/loras/opus100/az/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c472bc600cf404bb1e9096910456ddb6f1f5407e5b0199674499c0816c9b4cc +size 342547 diff --git a/loras/opus100/be/adapter_config.json b/loras/opus100/be/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/be/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/be/head_config.json b/loras/opus100/be/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/be/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/be/pytorch_adapter.bin b/loras/opus100/be/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1241b6faa0c8660be3cda3da0e15c3fd8aae8891 --- /dev/null +++ b/loras/opus100/be/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48caf1a38cc1b4753a9376a88f8c438f2e1615a05c2f7e4d70521abacf007042 +size 1333391 diff --git a/loras/opus100/be/pytorch_model_head.bin b/loras/opus100/be/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..810f002e206a8300e459d3d7400ce0315d8db739 --- /dev/null +++ b/loras/opus100/be/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:17a548ee8cbd7ab7c05be473dad1a3d9a0d929bdc3c15fc974fbda8e090a3ab9 +size 342547 diff --git a/loras/opus100/bg/adapter_config.json b/loras/opus100/bg/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/bg/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/bg/head_config.json b/loras/opus100/bg/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/bg/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/bg/pytorch_adapter.bin b/loras/opus100/bg/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..b2c2ed02923e78a865fe5d886e898d5f683f8da3 --- /dev/null +++ b/loras/opus100/bg/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6dc766c0473ca65e46837a36bdb8db0946185386537743f02e035f7dbf29c92 +size 1333391 diff --git a/loras/opus100/bg/pytorch_model_head.bin b/loras/opus100/bg/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..b8c43cb8f1b864c8cb393ee2b4494ecbd4d497db --- /dev/null +++ b/loras/opus100/bg/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a58ef254b4d6cbd604aba990e33ca596a0cfcf8f715ab42d862faa1f154f72b8 +size 342547 diff --git a/loras/opus100/bn/adapter_config.json b/loras/opus100/bn/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/bn/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/bn/head_config.json b/loras/opus100/bn/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/bn/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/bn/pytorch_adapter.bin b/loras/opus100/bn/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..4b1bb918bdda0adbb6b1d0ab9c87c3f8b0b8a05e --- /dev/null +++ b/loras/opus100/bn/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dbe875e8dfd86a16f80b563a8656046a70e1732e249585dbb3d6abcefab5466f +size 1333391 diff --git a/loras/opus100/bn/pytorch_model_head.bin b/loras/opus100/bn/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..ca76174a65bb59a7e630f3220c0828b6d87c650d --- /dev/null +++ b/loras/opus100/bn/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef926fb5ad255c90bdb92f2020f6f4f9957bebffe306c665b220d449345c16ee +size 342547 diff --git a/loras/opus100/ca/adapter_config.json b/loras/opus100/ca/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ca/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ca/head_config.json b/loras/opus100/ca/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ca/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ca/pytorch_adapter.bin b/loras/opus100/ca/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..e57dd606a36b248bd1066c7cab2fca4708cbb768 --- /dev/null +++ b/loras/opus100/ca/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:491eb19969e502f953627afa0e01f3fa9d40c7d4fc22c70f95ad65af3ecea029 +size 1333391 diff --git a/loras/opus100/ca/pytorch_model_head.bin b/loras/opus100/ca/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..1a3c9f077fb3cf64b3e46bb4929f868adeee121c --- /dev/null +++ b/loras/opus100/ca/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73b094821f486478edba9d53957ff5db7b5d995079d0cc0760aa90cd4247e9e4 +size 342547 diff --git a/loras/opus100/cs/adapter_config.json b/loras/opus100/cs/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/cs/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/cs/head_config.json b/loras/opus100/cs/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/cs/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/cs/pytorch_adapter.bin b/loras/opus100/cs/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..065ca353c4e9ac42698bf9b1be9eb653aeefea24 --- /dev/null +++ b/loras/opus100/cs/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1bf7e34edb1f93540be66d15ce72bf797b1d67f3f3145dfd878df2200179d7a +size 1333391 diff --git a/loras/opus100/cs/pytorch_model_head.bin b/loras/opus100/cs/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8b7b43bc0aaf46c5458c60849d0657fc8f7f34f4 --- /dev/null +++ b/loras/opus100/cs/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e11d140c03987f45edc499942c91f145e6938c221d8532c74ef50295adc4fe2 +size 342547 diff --git a/loras/opus100/cy/adapter_config.json b/loras/opus100/cy/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/cy/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/cy/head_config.json b/loras/opus100/cy/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/cy/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/cy/pytorch_adapter.bin b/loras/opus100/cy/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..5d27e228b445f4bf14ac5ee9801064b458da5d52 --- /dev/null +++ b/loras/opus100/cy/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:540b96255fc72ed5f7d739d0ec7a0298aa3e54fedf565da90f794b9b1f905c3b +size 1333391 diff --git a/loras/opus100/cy/pytorch_model_head.bin b/loras/opus100/cy/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..44325f9f73cebef4962409441f939e7978c5ac89 --- /dev/null +++ b/loras/opus100/cy/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:923931fa1ed759f7676d9479782bb53f9281646dfda578cba0b5282620ed56df +size 342547 diff --git a/loras/opus100/da/adapter_config.json b/loras/opus100/da/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/da/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/da/head_config.json b/loras/opus100/da/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/da/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/da/pytorch_adapter.bin b/loras/opus100/da/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..2bf14911fea98d13afc2d52c061ce69efc4bc303 --- /dev/null +++ b/loras/opus100/da/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b16c96a2a62c6936618691f30069ef583a941e5385807c56598eb2751d817454 +size 1333391 diff --git a/loras/opus100/da/pytorch_model_head.bin b/loras/opus100/da/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..ce4ccde57118aaf6e9365102c0ac5cd402d2bf2b --- /dev/null +++ b/loras/opus100/da/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d1a7338f2e2835b44fdb2e298e79cdd654931a8c1343f6191cf930fd143a782 +size 342547 diff --git a/loras/opus100/de/adapter_config.json b/loras/opus100/de/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/de/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/de/head_config.json b/loras/opus100/de/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/de/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/de/pytorch_adapter.bin b/loras/opus100/de/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..fd6b2ded4a11d27d50c39bc524a05346b6b8231a --- /dev/null +++ b/loras/opus100/de/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76e6186c1e284adbd42cd9fa5b27d0a3a179c73c36272f2ac6515a2efdfc67db +size 1333391 diff --git a/loras/opus100/de/pytorch_model_head.bin b/loras/opus100/de/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8682dbc8d3c9562d0950a79f0156899f179b53da --- /dev/null +++ b/loras/opus100/de/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27ea4547948edebadbd73650a44318e32b3c641847fe7d8a145b082529edfe8d +size 342547 diff --git a/loras/opus100/el/adapter_config.json b/loras/opus100/el/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/el/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/el/head_config.json b/loras/opus100/el/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/el/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/el/pytorch_adapter.bin b/loras/opus100/el/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..92a4ff137939df9c86bc8fbc018b449508ae45ef --- /dev/null +++ b/loras/opus100/el/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6f57f10579624d609ea2a00db4ad1906675f20f21c809203cf5f62c9ccdfe10 +size 1333391 diff --git a/loras/opus100/el/pytorch_model_head.bin b/loras/opus100/el/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..cb2ca5cfb0701a5a22c4c68df9855f2bfc084646 --- /dev/null +++ b/loras/opus100/el/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0fa09877656e2892283bd05370ed9f9a5650cec01f4c53bb0d02af9de68c94f +size 342547 diff --git a/loras/opus100/en/adapter_config.json b/loras/opus100/en/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/en/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/en/head_config.json b/loras/opus100/en/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/en/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/en/pytorch_adapter.bin b/loras/opus100/en/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..841f089c35a27e29f6c6a5091833d392acbcb834 --- /dev/null +++ b/loras/opus100/en/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fad6032b1d6dddda6f335720668aaca60519e17e9308df2d85598a2635315b28 +size 1333391 diff --git a/loras/opus100/en/pytorch_model_head.bin b/loras/opus100/en/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..964add01dc0bd8d444d0ca89932df3d5079443bb --- /dev/null +++ b/loras/opus100/en/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b4c8d835108a413d18b374bb718bfecd4a4d39cc61d1da31a84236a4be1a1fb +size 342547 diff --git a/loras/opus100/eo/adapter_config.json b/loras/opus100/eo/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/eo/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/eo/head_config.json b/loras/opus100/eo/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/eo/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/eo/pytorch_adapter.bin b/loras/opus100/eo/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..4db5c1b498d1465b9b12789a9987b35d07eb07df --- /dev/null +++ b/loras/opus100/eo/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f2860065104a3828d311d22799ef81e7ea90a159d65ae140b5bd38027810712 +size 1333391 diff --git a/loras/opus100/eo/pytorch_model_head.bin b/loras/opus100/eo/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..f119c1caa7962c1a3354a7f280533c86637657ff --- /dev/null +++ b/loras/opus100/eo/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d8f4a37983fa4ec59d2b19b64d47421c212cdf9f3b5f97de3a90b4b37759313 +size 342547 diff --git a/loras/opus100/es/adapter_config.json b/loras/opus100/es/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/es/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/es/head_config.json b/loras/opus100/es/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/es/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/es/pytorch_adapter.bin b/loras/opus100/es/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..eec13cae8d57843e28c8fe82596c2e9b9c792457 --- /dev/null +++ b/loras/opus100/es/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c657e0c80b0907c42965ec26bf3ceaf9514f1d7d672b807da59cf4033a8b9448 +size 1333391 diff --git a/loras/opus100/es/pytorch_model_head.bin b/loras/opus100/es/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..4f18ae2fcafbf66b916a127b83380649ba032f23 --- /dev/null +++ b/loras/opus100/es/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c3a1443c01f2d643bd3a7afe584c77f8cf19f5bc69db6a29b6a081bd521c9a8 +size 342547 diff --git a/loras/opus100/et/adapter_config.json b/loras/opus100/et/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/et/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/et/head_config.json b/loras/opus100/et/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/et/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/et/pytorch_adapter.bin b/loras/opus100/et/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..58a403cb29e9786a05546f61407d8a7d695c81c2 --- /dev/null +++ b/loras/opus100/et/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57b0de6c19e77fe3f8ac8c1e2f3fb6fa06a5ba6d6af3ed2441dfefc96e8032f6 +size 1333391 diff --git a/loras/opus100/et/pytorch_model_head.bin b/loras/opus100/et/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..c8c4a7436548e8de81faa9421deff7445e9f5071 --- /dev/null +++ b/loras/opus100/et/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fdcf6e850f6893d313bafc1e87baa7bcceabffdb7547a1bb0b13a0a9ba919c2d +size 342547 diff --git a/loras/opus100/eu/adapter_config.json b/loras/opus100/eu/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/eu/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/eu/head_config.json b/loras/opus100/eu/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/eu/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/eu/pytorch_adapter.bin b/loras/opus100/eu/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..67370c12fdf796c08915e3bc7ea12daea1f4fbb8 --- /dev/null +++ b/loras/opus100/eu/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61098553ea1b8c68bbeeb7f99bbde019154a7f1fb3b61c34f46e2d81b3fb3847 +size 1333391 diff --git a/loras/opus100/eu/pytorch_model_head.bin b/loras/opus100/eu/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..51f16c556b98d60d74295d567ef3bb3c6ed49438 --- /dev/null +++ b/loras/opus100/eu/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3cfca07bd43d9a4c97abbced56eafc112c3e0969e9414082aff6eebce64d600f +size 342547 diff --git a/loras/opus100/fa/adapter_config.json b/loras/opus100/fa/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/fa/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/fa/head_config.json b/loras/opus100/fa/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/fa/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/fa/pytorch_adapter.bin b/loras/opus100/fa/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1983ff277ba3e21c606d758b85d0e588bf474aed --- /dev/null +++ b/loras/opus100/fa/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:adae22d4c8350f5db38e694324a29310038769cfd6dcd45cb3c9931a5590cc06 +size 1333391 diff --git a/loras/opus100/fa/pytorch_model_head.bin b/loras/opus100/fa/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..0556070f85bb80a4ae5eb2662d1039b64f8e3fbd --- /dev/null +++ b/loras/opus100/fa/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f117ef0471773eebf0e11a8da0bd3603b791ed67ef62955dcaf9ede68c2e58aa +size 342547 diff --git a/loras/opus100/fi/adapter_config.json b/loras/opus100/fi/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/fi/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/fi/head_config.json b/loras/opus100/fi/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/fi/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/fi/pytorch_adapter.bin b/loras/opus100/fi/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..2521d7ca9c2ae19bdf2a858ab62f05839d096084 --- /dev/null +++ b/loras/opus100/fi/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa16af57fe9d24a16ba9719158929d1d18149e11f211c02fbe7792b62170472c +size 1333391 diff --git a/loras/opus100/fi/pytorch_model_head.bin b/loras/opus100/fi/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..ca7a06556c5d484339a2a9b75ba65810dcb84725 --- /dev/null +++ b/loras/opus100/fi/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ad6cf6b9ee5d90d1246a273f75694a0d84c48e89cb11a8c1d6771ded8be72d7 +size 342547 diff --git a/loras/opus100/fy/adapter_config.json b/loras/opus100/fy/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/fy/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/fy/head_config.json b/loras/opus100/fy/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/fy/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/fy/pytorch_adapter.bin b/loras/opus100/fy/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..f72eb4d109f5c906c5509c85eeabe6732e4548a3 --- /dev/null +++ b/loras/opus100/fy/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3596d7bcf4bee7d21dbef84efe0ca767037519f0eb6554e100d52e7579280f58 +size 1333391 diff --git a/loras/opus100/fy/pytorch_model_head.bin b/loras/opus100/fy/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..778805380eaea44c85e7cedd28563613803a2a59 --- /dev/null +++ b/loras/opus100/fy/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10b6bb388e5fb13b7f6c111c845efc2657af71d6b67d27884da697ab2d8a8460 +size 342547 diff --git a/loras/opus100/ga/adapter_config.json b/loras/opus100/ga/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ga/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ga/head_config.json b/loras/opus100/ga/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ga/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ga/pytorch_adapter.bin b/loras/opus100/ga/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..3ec932dd4b922a64709b02ad44f1c3f112d6bdd8 --- /dev/null +++ b/loras/opus100/ga/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10e4e7651201da44c182573416e0f936b44bf50b731f0ea012221a0631a35414 +size 1333391 diff --git a/loras/opus100/ga/pytorch_model_head.bin b/loras/opus100/ga/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..6c816cd221e3c5f9822bde0163bddd75025204e4 --- /dev/null +++ b/loras/opus100/ga/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27e331763726ecf61a8e43c8cd0788d0b56e60c31b53ab1e732429a5100c5ba3 +size 342547 diff --git a/loras/opus100/gd/adapter_config.json b/loras/opus100/gd/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/gd/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/gd/head_config.json b/loras/opus100/gd/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/gd/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/gd/pytorch_adapter.bin b/loras/opus100/gd/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..580bf925825ad09159d6d430a1e40d9922f78227 --- /dev/null +++ b/loras/opus100/gd/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aaeaabba32eb34aa0f1dd6b0127742208529246bbb130bc13fd6bf78ecb86d7f +size 1333391 diff --git a/loras/opus100/gd/pytorch_model_head.bin b/loras/opus100/gd/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..093eb32a0b144130d7faff73e05f5d3f9da0b2e7 --- /dev/null +++ b/loras/opus100/gd/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31441811e7760a06b2f4fa6e78acb02fcb803fba27724b1f1552c2aaee92104d +size 342547 diff --git a/loras/opus100/gl/adapter_config.json b/loras/opus100/gl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/gl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/gl/head_config.json b/loras/opus100/gl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/gl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/gl/pytorch_adapter.bin b/loras/opus100/gl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..5229d77318a5427e1efd13722d25e78f0e2e5b72 --- /dev/null +++ b/loras/opus100/gl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:160a09c05dd521cf37e799917252b289cbd23b7a99f2dccc41785ba6e782f8df +size 1333391 diff --git a/loras/opus100/gl/pytorch_model_head.bin b/loras/opus100/gl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..89d19de3260e7907473205fb8466c5392f4b545a --- /dev/null +++ b/loras/opus100/gl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a747204b7edf37c9bf90f8c40a95c69b86fe95de7e4e4ab446d035d63f8ce13 +size 342547 diff --git a/loras/opus100/gu/adapter_config.json b/loras/opus100/gu/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/gu/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/gu/head_config.json b/loras/opus100/gu/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/gu/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/gu/pytorch_adapter.bin b/loras/opus100/gu/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..355438897cb25318190c33610d969a545cb35882 --- /dev/null +++ b/loras/opus100/gu/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b288a2622aeba3a887d5aa16b7ecd964fbeafcde2ab7a39b745aa1a21c0c09b0 +size 1333391 diff --git a/loras/opus100/gu/pytorch_model_head.bin b/loras/opus100/gu/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..4480d7dc9f1833a27bfc1b2d8be9c6f15d4129d1 --- /dev/null +++ b/loras/opus100/gu/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:027119395173a91228b353f204e6f385cd8f744c6169dc32f4eb53d8e9e3bb14 +size 342547 diff --git a/loras/opus100/ha/adapter_config.json b/loras/opus100/ha/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ha/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ha/head_config.json b/loras/opus100/ha/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ha/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ha/pytorch_adapter.bin b/loras/opus100/ha/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..470b5574b5114c98faf08d625d4cfd52c60d8040 --- /dev/null +++ b/loras/opus100/ha/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff7f835acd09a1aaa8d62de535e34367e3d11b43a65623e1975b51a59599ca4e +size 1333391 diff --git a/loras/opus100/ha/pytorch_model_head.bin b/loras/opus100/ha/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..a69fc655094117a69fd18058350f46799e504442 --- /dev/null +++ b/loras/opus100/ha/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35d6e094d0f6c08a7067cfee77e0bd6bca1bdb55064bb6a195f2738c598daa48 +size 342547 diff --git a/loras/opus100/he/adapter_config.json b/loras/opus100/he/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/he/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/he/head_config.json b/loras/opus100/he/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/he/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/he/pytorch_adapter.bin b/loras/opus100/he/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d487260ab00055d24cc9c95042612901c5f28ee2 --- /dev/null +++ b/loras/opus100/he/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df83d16388e62c283d6260339709c42baf74a42dc0fd948f717851f13d431ad0 +size 1333391 diff --git a/loras/opus100/he/pytorch_model_head.bin b/loras/opus100/he/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..36278e798235794f7168da3b50d11093366053eb --- /dev/null +++ b/loras/opus100/he/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74c816966fdadd8e33ac6915654e87776886f5b2a3fdaf99c7558d6aef4a4d61 +size 342547 diff --git a/loras/opus100/hi/adapter_config.json b/loras/opus100/hi/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/hi/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/hi/head_config.json b/loras/opus100/hi/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/hi/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/hi/pytorch_adapter.bin b/loras/opus100/hi/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..dc39765d4bd9ece27aff091a7f16f2c7bf279821 --- /dev/null +++ b/loras/opus100/hi/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1adae447c8c58ffed40cbb4b72cf1614867f0037fe885bc8e8040b7d4284fead +size 1333391 diff --git a/loras/opus100/hi/pytorch_model_head.bin b/loras/opus100/hi/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..65354c73952e7df271f49843b5ce494514d0e944 --- /dev/null +++ b/loras/opus100/hi/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0876715cb71b0988214f6dcb9db368f0cbf27f5fc8a3c79d1cd0e9da1ea84482 +size 342547 diff --git a/loras/opus100/hu/adapter_config.json b/loras/opus100/hu/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/hu/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/hu/head_config.json b/loras/opus100/hu/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/hu/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/hu/pytorch_adapter.bin b/loras/opus100/hu/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..005d73b6aa9b51c96982fc9cea906e6b2dec29f2 --- /dev/null +++ b/loras/opus100/hu/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2c76459d926e2c41132cb8982b63e2ec800218ea4fc77efddaa0b5165a7fcff +size 1333391 diff --git a/loras/opus100/hu/pytorch_model_head.bin b/loras/opus100/hu/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8681fd2a9b424b3406cfe675b27b2117e38022ca --- /dev/null +++ b/loras/opus100/hu/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10f7e977f8d95ddf1ee3712d3da773b31e32f30050a46ab94a617b3bb1e7446a +size 342547 diff --git a/loras/opus100/id/adapter_config.json b/loras/opus100/id/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/id/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/id/head_config.json b/loras/opus100/id/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/id/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/id/pytorch_adapter.bin b/loras/opus100/id/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ac89ec5c4b97e8213c6c668ad6f422c4eec578cb --- /dev/null +++ b/loras/opus100/id/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:009fc9f03578a17184fb8a52eaa9a1fa98ee584336b152af47268dab9251974f +size 1333391 diff --git a/loras/opus100/id/pytorch_model_head.bin b/loras/opus100/id/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..5af505c9bc42f096b0207e3b3cd1d1f1a965cb7a --- /dev/null +++ b/loras/opus100/id/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67ba93ca7a4f5fa5b6bf785973c3d335d86c86e8a05ab5d9586a89267b8e6628 +size 342547 diff --git a/loras/opus100/ig/adapter_config.json b/loras/opus100/ig/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ig/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ig/head_config.json b/loras/opus100/ig/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ig/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ig/pytorch_adapter.bin b/loras/opus100/ig/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..6dbee5819b39ef609e0a6da0548f269e4fbd67b1 --- /dev/null +++ b/loras/opus100/ig/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1dca7a97a56efdf08b88eeb0d3c1e30f841452726dc233294851e5f9469538a1 +size 1333391 diff --git a/loras/opus100/ig/pytorch_model_head.bin b/loras/opus100/ig/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..26f3f33fd92c5b1b308629516ab9967cd3e88e8d --- /dev/null +++ b/loras/opus100/ig/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:baf421f26043248d39864a54f3243989205b1d68e1774552372a3d02915ce676 +size 342547 diff --git a/loras/opus100/is/adapter_config.json b/loras/opus100/is/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/is/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/is/head_config.json b/loras/opus100/is/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/is/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/is/pytorch_adapter.bin b/loras/opus100/is/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1fc3cd040e6fef431c5ef5f2f6dd6800c8f8268d --- /dev/null +++ b/loras/opus100/is/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd99372e6f221a10b241907d8f502416a5e5cf57cdda9f36e45983737b7b9fe3 +size 1333391 diff --git a/loras/opus100/is/pytorch_model_head.bin b/loras/opus100/is/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8fb4103f14dd8246c260908efcde6bc5b8c9cb5b --- /dev/null +++ b/loras/opus100/is/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7a38868ce58d57bd46101409c9141b052977e0cdf601c22b32af170e784b78f +size 342547 diff --git a/loras/opus100/it/adapter_config.json b/loras/opus100/it/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/it/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/it/head_config.json b/loras/opus100/it/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/it/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/it/pytorch_adapter.bin b/loras/opus100/it/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..c41638abe460c21f32e4f959b884d4459d16c92e --- /dev/null +++ b/loras/opus100/it/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc58584b2ef8e998cdee16ba0103dc340157d6c3982e33f7d846b7700be7737b +size 1333391 diff --git a/loras/opus100/it/pytorch_model_head.bin b/loras/opus100/it/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..b4aca896af19cdaa70e560233cd028b014323e2e --- /dev/null +++ b/loras/opus100/it/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:982979bfa5efe2850322dbfffe3754c35692f5cf1c50e3aab95ea3b53c5ce680 +size 342547 diff --git a/loras/opus100/ja/adapter_config.json b/loras/opus100/ja/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ja/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ja/head_config.json b/loras/opus100/ja/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ja/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ja/pytorch_adapter.bin b/loras/opus100/ja/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1704400d28fd8426428d1c12661f89ebc6dd5ea1 --- /dev/null +++ b/loras/opus100/ja/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c80704a6a5bd8d71c4345b968fd563dab2b7bb9562570670a529b12a50e63a98 +size 1333391 diff --git a/loras/opus100/ja/pytorch_model_head.bin b/loras/opus100/ja/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..d1a7968326d098ca232930d76fb3953f39f7f751 --- /dev/null +++ b/loras/opus100/ja/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7c856946e7bef143d8ed39754c3790897e9753b742514f3e9279bb8736c3a8f +size 342547 diff --git a/loras/opus100/ka/adapter_config.json b/loras/opus100/ka/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ka/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ka/head_config.json b/loras/opus100/ka/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ka/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ka/pytorch_adapter.bin b/loras/opus100/ka/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..fedd2b38ed63449f38257437345e083670953d35 --- /dev/null +++ b/loras/opus100/ka/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6ca02335f013897856469488bdc17ee92ca9b01f9aac66b4df2d3da09fcae83 +size 1333391 diff --git a/loras/opus100/ka/pytorch_model_head.bin b/loras/opus100/ka/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..591315bcc2fb0d16c06723ec509f44870be28ea7 --- /dev/null +++ b/loras/opus100/ka/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee0d31a665418201875f0c010f5a7aa63a0a05246dc7539c7399ebe93981f3c0 +size 342547 diff --git a/loras/opus100/kk/adapter_config.json b/loras/opus100/kk/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/kk/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/kk/head_config.json b/loras/opus100/kk/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/kk/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/kk/pytorch_adapter.bin b/loras/opus100/kk/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..5b5a51fa2de66c18acd9b66f2a9374588bdc4064 --- /dev/null +++ b/loras/opus100/kk/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ecad7b78c006c7af2eda59bb1e8d49c752817ef56075191bd0136f09aace59f +size 1333391 diff --git a/loras/opus100/kk/pytorch_model_head.bin b/loras/opus100/kk/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..f3677fd1c7783eb7d13eb07e6f463752f6a2a6a3 --- /dev/null +++ b/loras/opus100/kk/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0d343ef1577169ba4bd553d26a1edebcb9cdce0f45cda0f32c771c6a54f4709 +size 342547 diff --git a/loras/opus100/km/adapter_config.json b/loras/opus100/km/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/km/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/km/head_config.json b/loras/opus100/km/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/km/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/km/pytorch_adapter.bin b/loras/opus100/km/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..bd75c5f539615def651d3689344f792f382cfe82 --- /dev/null +++ b/loras/opus100/km/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19b73bf6fd8b88e89274926e975fd48ed4c5c06bf084afa06d353f8b398a85c5 +size 1333391 diff --git a/loras/opus100/km/pytorch_model_head.bin b/loras/opus100/km/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..18f463b87a228a1c1b6df14bb73cfd3ebd7e1158 --- /dev/null +++ b/loras/opus100/km/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c6eb420ce39c424eaeb193596b41bd8cb13029c73d4c7246bca5452ddc3e58c +size 342547 diff --git a/loras/opus100/kn/adapter_config.json b/loras/opus100/kn/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/kn/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/kn/head_config.json b/loras/opus100/kn/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/kn/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/kn/pytorch_adapter.bin b/loras/opus100/kn/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..7f37210c255ec8b2819cc5ac83ff1b2695f61b0f --- /dev/null +++ b/loras/opus100/kn/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2267a12ef84605dde8618a9d8780069f37e7f46807873168e29c77ee4563077 +size 1333391 diff --git a/loras/opus100/kn/pytorch_model_head.bin b/loras/opus100/kn/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..9a4d8f0b2fc7994145bee7f6d31a94e2d2d37453 --- /dev/null +++ b/loras/opus100/kn/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7051070dc00f23e4bafc424bb681e75e939c5b3b50a3d8a5e878c364bb546ac +size 342547 diff --git a/loras/opus100/ko/adapter_config.json b/loras/opus100/ko/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ko/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ko/head_config.json b/loras/opus100/ko/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ko/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ko/pytorch_adapter.bin b/loras/opus100/ko/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1852c96e7cb39d50e7f6a2eb5914c22629288160 --- /dev/null +++ b/loras/opus100/ko/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3eeddb343358b93634562eb258f04c5bfcf72169f1ad8bfc8b6fc6577e5c06d0 +size 1333391 diff --git a/loras/opus100/ko/pytorch_model_head.bin b/loras/opus100/ko/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..c041adb54b1ff594ede89469977c05f30473618c --- /dev/null +++ b/loras/opus100/ko/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f24341739cd95d82e4902033ab352eec3308c505eb0976cc0e0987914cc2c836 +size 342547 diff --git a/loras/opus100/ku/adapter_config.json b/loras/opus100/ku/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ku/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ku/head_config.json b/loras/opus100/ku/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ku/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ku/pytorch_adapter.bin b/loras/opus100/ku/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..7aa037fb6a878232c6d35dac0e052dc8ccbba12a --- /dev/null +++ b/loras/opus100/ku/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44dc61e8ad31bcb967ea2f6ba5ba7be1a8b3d25a7dd4332c3e0c73724d75adf6 +size 1333391 diff --git a/loras/opus100/ku/pytorch_model_head.bin b/loras/opus100/ku/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..6b9db2d1f6ba40b3ca839fa94a6515254f28d5a4 --- /dev/null +++ b/loras/opus100/ku/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56d0b5954304abd8140a9a5e8e740620d486dbd2155a52cb6ea074a348144ba8 +size 342547 diff --git a/loras/opus100/ky/adapter_config.json b/loras/opus100/ky/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ky/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ky/head_config.json b/loras/opus100/ky/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ky/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ky/pytorch_adapter.bin b/loras/opus100/ky/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1bdbc220c97fd2a8973fc499c6fa87ca9706bfe8 --- /dev/null +++ b/loras/opus100/ky/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72d9e8c8411fe524c16be00b7810b50e5e82f382f5d06825dd448f0d18d3caf1 +size 1333391 diff --git a/loras/opus100/ky/pytorch_model_head.bin b/loras/opus100/ky/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..7fc7d49c432ac2a96f48612a52b005a5c7c81e87 --- /dev/null +++ b/loras/opus100/ky/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03e4ef54c8975a69e761d8e3b15ed416248939b46c7465bd3d86a94ceb061f5c +size 342547 diff --git a/loras/opus100/lt/adapter_config.json b/loras/opus100/lt/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/lt/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/lt/head_config.json b/loras/opus100/lt/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/lt/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/lt/pytorch_adapter.bin b/loras/opus100/lt/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..654465c444247b5ba2f4002e3fda7059f8a8d767 --- /dev/null +++ b/loras/opus100/lt/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb1e91e19fdbc8250903a166051937e871a32d20339845f25220f870ae741b3d +size 1333391 diff --git a/loras/opus100/lt/pytorch_model_head.bin b/loras/opus100/lt/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..883f6c42ba81c2231c4ee3d8524ab913a2d26e71 --- /dev/null +++ b/loras/opus100/lt/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6bbeb55dd485463732e37875799f0bf69169901ba1ec2e6268e249327b3e788e +size 342547 diff --git a/loras/opus100/lv/adapter_config.json b/loras/opus100/lv/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/lv/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/lv/head_config.json b/loras/opus100/lv/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/lv/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/lv/pytorch_adapter.bin b/loras/opus100/lv/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..2fb81b56205ebed47204a2ad7656a150aed47596 --- /dev/null +++ b/loras/opus100/lv/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b4ab5d92dd0077a41abbacd8d7f42a266d5e6660b4dcd9b2b52206dd263d1c8 +size 1333391 diff --git a/loras/opus100/lv/pytorch_model_head.bin b/loras/opus100/lv/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..c68c93b06542ed19dc7b08b0b2ea5fc6087691bd --- /dev/null +++ b/loras/opus100/lv/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a77b2ff8717af1ee83fea03cdec741c7599c4661772b091fee672d43635ba7a +size 342547 diff --git a/loras/opus100/mg/adapter_config.json b/loras/opus100/mg/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/mg/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/mg/head_config.json b/loras/opus100/mg/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/mg/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/mg/pytorch_adapter.bin b/loras/opus100/mg/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..30e56cd93760375039fd82cf9274f41320e49b08 --- /dev/null +++ b/loras/opus100/mg/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e2a65acbb33ce50620a5699c07ce4443b04f18624f5787d27d0ec28dbea9ef6 +size 1333391 diff --git a/loras/opus100/mg/pytorch_model_head.bin b/loras/opus100/mg/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..865d277c2d0e22637030382ff126e2bea12562e6 --- /dev/null +++ b/loras/opus100/mg/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33cad2df0f445e5b2775836073464719d711c5c66f70244dd51cb96675dfca63 +size 342547 diff --git a/loras/opus100/mk/adapter_config.json b/loras/opus100/mk/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/mk/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/mk/head_config.json b/loras/opus100/mk/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/mk/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/mk/pytorch_adapter.bin b/loras/opus100/mk/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..05905be8384e7ff672aa14020f2662becfd66e8d --- /dev/null +++ b/loras/opus100/mk/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d598419fb0e30827a2cfb18f90f1be6136877f695aa1fcbe3bef1e958d795629 +size 1333391 diff --git a/loras/opus100/mk/pytorch_model_head.bin b/loras/opus100/mk/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..17df0dfce25a25baca90a3bc7506aef925446566 --- /dev/null +++ b/loras/opus100/mk/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0569647e8b00e790e66fd0d4c1ec7f0e2fa00b6ad6df128829fec1792d0c0b0 +size 342547 diff --git a/loras/opus100/ml/adapter_config.json b/loras/opus100/ml/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ml/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ml/head_config.json b/loras/opus100/ml/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ml/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ml/pytorch_adapter.bin b/loras/opus100/ml/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..e53537b9e2f34d03909203339b33617113c6d581 --- /dev/null +++ b/loras/opus100/ml/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f99921e1345c7741af9282f3cae4b9372c1957056e04fd15ed09b8433c267a2 +size 1333391 diff --git a/loras/opus100/ml/pytorch_model_head.bin b/loras/opus100/ml/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..e270f60b7853e6489ae70fd9343a054cefab9a5b --- /dev/null +++ b/loras/opus100/ml/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:169dcaf36746c8e8bf4b39a39c96b863b4491ed5a4b1113b13eefdfd6d893471 +size 342547 diff --git a/loras/opus100/mr/adapter_config.json b/loras/opus100/mr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/mr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/mr/head_config.json b/loras/opus100/mr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/mr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/mr/pytorch_adapter.bin b/loras/opus100/mr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..8e57c3660b2ffe3ab17c7f12dcee538b3b417a04 --- /dev/null +++ b/loras/opus100/mr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e5b9187b156e59d3f8095bdfe0665a64cd8b64a7b8fcb3515cc4bc74f19e51f +size 1333391 diff --git a/loras/opus100/mr/pytorch_model_head.bin b/loras/opus100/mr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..1bc1851e36556c56941dfc3f3ca3c72b6726e34f --- /dev/null +++ b/loras/opus100/mr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bdf564ac15d9fcf5c0d777722f4a5013545cbf95cfbb638c68ad4ac13b0a9e27 +size 342547 diff --git a/loras/opus100/ms/adapter_config.json b/loras/opus100/ms/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ms/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ms/head_config.json b/loras/opus100/ms/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ms/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ms/pytorch_adapter.bin b/loras/opus100/ms/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..2af479c4c4beaf6a0ad26fe78a2d4fb257a1b2f9 --- /dev/null +++ b/loras/opus100/ms/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e23139962bbd98ca3e760a87e9c326d454531dc7b2d64f64c13794180d2d099 +size 1333391 diff --git a/loras/opus100/ms/pytorch_model_head.bin b/loras/opus100/ms/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8b22d52cda5a00339125cbe99448bed6d0f3449f --- /dev/null +++ b/loras/opus100/ms/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1fbe11c3d92b84ae6eab6217c48235e76caa17d3024c30051e1d191b378597f0 +size 342547 diff --git a/loras/opus100/mt/adapter_config.json b/loras/opus100/mt/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/mt/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/mt/head_config.json b/loras/opus100/mt/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/mt/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/mt/pytorch_adapter.bin b/loras/opus100/mt/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..bdbd10baf2d24e801f8396ca28f11a4f6ae51e9f --- /dev/null +++ b/loras/opus100/mt/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:816c0fbbac67424f920433fee6f9f410c5791a95f23494e47d5793123452327b +size 1333391 diff --git a/loras/opus100/mt/pytorch_model_head.bin b/loras/opus100/mt/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..6cd3f9d7a2e3f83a6591aaebdf9ceaf18c072cf8 --- /dev/null +++ b/loras/opus100/mt/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c37baaaa1d9ce1e84825f4ae0443ba15489bc7e3511417e918caf5aa34cd6ca9 +size 342547 diff --git a/loras/opus100/my/adapter_config.json b/loras/opus100/my/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/my/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/my/head_config.json b/loras/opus100/my/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/my/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/my/pytorch_adapter.bin b/loras/opus100/my/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ca0013851c1a6c6e67b4f90cf7fb90226771c3b5 --- /dev/null +++ b/loras/opus100/my/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1fdd42159654559946b25f8836a32129c37ae3220f0013094c446c30c6202d28 +size 1333391 diff --git a/loras/opus100/my/pytorch_model_head.bin b/loras/opus100/my/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..a642ca7c5c91c2007f7f1192af4c3b99f9b2e345 --- /dev/null +++ b/loras/opus100/my/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e98f5e9c75f5f6a7a0ebd061baa64ee233a641eaf168ea26be17139e9087817c +size 342547 diff --git a/loras/opus100/ne/adapter_config.json b/loras/opus100/ne/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ne/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ne/head_config.json b/loras/opus100/ne/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ne/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ne/pytorch_adapter.bin b/loras/opus100/ne/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..2b60ca3ec33e5c1d92bdd9fc93ff5ac05e002357 --- /dev/null +++ b/loras/opus100/ne/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05fe5ecedbb3ce4976c3c07c419a89b8788603510ad4374f7ae3983079407a1f +size 1333391 diff --git a/loras/opus100/ne/pytorch_model_head.bin b/loras/opus100/ne/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..a20091c65dd53af0f22ea7e890e0bc0bf077149e --- /dev/null +++ b/loras/opus100/ne/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1711664ab9e08b5d7f4b4d9d88235ec17b2c0f529600f6c07826157294638ae2 +size 342547 diff --git a/loras/opus100/no/adapter_config.json b/loras/opus100/no/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/no/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/no/head_config.json b/loras/opus100/no/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/no/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/no/pytorch_adapter.bin b/loras/opus100/no/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..26a69f3ccdd168801f6a767457aba04d9a0eb37a --- /dev/null +++ b/loras/opus100/no/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c0e193cf9f1f772f5acb498ae0d27991c4a701f404a88e4dde6798c204bcf53 +size 1333391 diff --git a/loras/opus100/no/pytorch_model_head.bin b/loras/opus100/no/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..4d951c3d45d82dac608368fd3b4006f02e0eaaa6 --- /dev/null +++ b/loras/opus100/no/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b29132344ae91714a8135b3e88c59022b78e5471a602e06362be9c36c8feb22 +size 342547 diff --git a/loras/opus100/pa/adapter_config.json b/loras/opus100/pa/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/pa/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/pa/head_config.json b/loras/opus100/pa/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/pa/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/pa/pytorch_adapter.bin b/loras/opus100/pa/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..159243beaeda00a4ae700dde2065b823f2634798 --- /dev/null +++ b/loras/opus100/pa/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5be31fa7e807264754ef8652e4af82d81e3f13ae0cd1f1e9f6ab6f708d167cda +size 1333391 diff --git a/loras/opus100/pa/pytorch_model_head.bin b/loras/opus100/pa/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..9917359f8affd9f3a16c61761fad9682146ef45f --- /dev/null +++ b/loras/opus100/pa/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10cc101e55009b1529f146c3586fae0c3f254560f1f790f43b4c32a184725ae4 +size 342547 diff --git a/loras/opus100/pl/adapter_config.json b/loras/opus100/pl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/pl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/pl/head_config.json b/loras/opus100/pl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/pl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/pl/pytorch_adapter.bin b/loras/opus100/pl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..03ab099058253fecf5e1aba8e9ac9f26a15f87ad --- /dev/null +++ b/loras/opus100/pl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f91ece50b50eec67c68ff9fdcbe15e6b1f09e1c8427d34b5d6d60be4ee20f982 +size 1333391 diff --git a/loras/opus100/pl/pytorch_model_head.bin b/loras/opus100/pl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..f3f1cf44ff0e2ffad71c76531de2512aca742597 --- /dev/null +++ b/loras/opus100/pl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30fdf32ae8d79edc99930f3bf4ac535b82c769cb3b1411869cb4fe0df3f9b063 +size 342547 diff --git a/loras/opus100/ps/adapter_config.json b/loras/opus100/ps/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ps/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ps/head_config.json b/loras/opus100/ps/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ps/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ps/pytorch_adapter.bin b/loras/opus100/ps/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..cb84d3c997b568c9179cae9b4cf812045b28a1cc --- /dev/null +++ b/loras/opus100/ps/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d0430a1401ee9616c46e98341c893e950043595cb9b92eddf0405e2e909b7364 +size 1333391 diff --git a/loras/opus100/ps/pytorch_model_head.bin b/loras/opus100/ps/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..a1f70c5b401d45e19481b2374a532edb34a40467 --- /dev/null +++ b/loras/opus100/ps/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84a903b9da485dec6f8e066cbe0f0d2810ec230f3b2b5b0995b571330ea569ec +size 342547 diff --git a/loras/opus100/pt/adapter_config.json b/loras/opus100/pt/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/pt/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/pt/head_config.json b/loras/opus100/pt/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/pt/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/pt/pytorch_adapter.bin b/loras/opus100/pt/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..7541347ddcdd366248f35c3e1b5ccd9cd6ec8639 --- /dev/null +++ b/loras/opus100/pt/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31f7f3fdfdf2ca31dfeedba0448cf0460e8db343b6b78f93a26566f85014372b +size 1333391 diff --git a/loras/opus100/pt/pytorch_model_head.bin b/loras/opus100/pt/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..fd4005a8a224d29739ccdd4cf14dde7d232a1159 --- /dev/null +++ b/loras/opus100/pt/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:368437cff5b9959c04f00520c2e3ba155cf9d747f6248c6984551e3162c27c62 +size 342547 diff --git a/loras/opus100/ro/adapter_config.json b/loras/opus100/ro/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ro/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ro/head_config.json b/loras/opus100/ro/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ro/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ro/pytorch_adapter.bin b/loras/opus100/ro/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..5d5ac1f653acc1a1e45ef2502b7d04e2b1015844 --- /dev/null +++ b/loras/opus100/ro/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:135ff54e805fee0e2bf8435418ea823c4b0a8db9b6e3852dadb27e3a5d5b5704 +size 1333391 diff --git a/loras/opus100/ro/pytorch_model_head.bin b/loras/opus100/ro/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..e962b405dd933da33ff73cc6c5ca2a9611782ef3 --- /dev/null +++ b/loras/opus100/ro/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:332c144c1f17551465d097320116a27a04279a738575ea4c8550bf5866071a7b +size 342547 diff --git a/loras/opus100/si/adapter_config.json b/loras/opus100/si/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/si/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/si/head_config.json b/loras/opus100/si/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/si/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/si/pytorch_adapter.bin b/loras/opus100/si/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d2ff09ecbf820ae6d0d2ab49fb2de49fecfed03f --- /dev/null +++ b/loras/opus100/si/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71307700d97b03871d25f184b37e8d42993939a048187bc7fd0ea75233645836 +size 1333391 diff --git a/loras/opus100/si/pytorch_model_head.bin b/loras/opus100/si/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..5e98f4f5c614472810a255bd65a2a8dfd0898c67 --- /dev/null +++ b/loras/opus100/si/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1d2c258823e9b9041f52431fba057a852a12187472c0561fca66f2b46c3df19 +size 342547 diff --git a/loras/opus100/sk/adapter_config.json b/loras/opus100/sk/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/sk/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/sk/head_config.json b/loras/opus100/sk/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/sk/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/sk/pytorch_adapter.bin b/loras/opus100/sk/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..5b7de5f6cf042b7217e25d90ccdcd3078146cb7c --- /dev/null +++ b/loras/opus100/sk/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31a07ad9bc116dc4e4da1a509174d4de68a9975a3e6aa8799cce3f0ebad26385 +size 1333391 diff --git a/loras/opus100/sk/pytorch_model_head.bin b/loras/opus100/sk/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..27414be5e1f4c802283cb6a9cdb01c954bc1c896 --- /dev/null +++ b/loras/opus100/sk/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34346677fcd90bbaeb5fa482c2d2426ba007cd4933e2cad918123a21a182e73e +size 342547 diff --git a/loras/opus100/sl/adapter_config.json b/loras/opus100/sl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/sl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/sl/head_config.json b/loras/opus100/sl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/sl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/sl/pytorch_adapter.bin b/loras/opus100/sl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..f12314c6f03e86eda05eac21b4bc6fcc9c7684a6 --- /dev/null +++ b/loras/opus100/sl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f3d06906c334218052df9eb0a8999f3db179735424d7ad589c0788794ca002d +size 1333391 diff --git a/loras/opus100/sl/pytorch_model_head.bin b/loras/opus100/sl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..2588724f40f1864eb228752733141601bc8ac7a1 --- /dev/null +++ b/loras/opus100/sl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c949a3b47c9704b54612aa1abda403bbcf29b29d7fa1652c08cce54bac1a7be1 +size 342547 diff --git a/loras/opus100/sq/adapter_config.json b/loras/opus100/sq/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/sq/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/sq/head_config.json b/loras/opus100/sq/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/sq/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/sq/pytorch_adapter.bin b/loras/opus100/sq/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..98044e87cb3fff095eb549d2122940a4e495d2bb --- /dev/null +++ b/loras/opus100/sq/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d365cf1405ee1fb37270d97f2e1b42603653c675c60c0b84402503309aaf3caf +size 1333391 diff --git a/loras/opus100/sq/pytorch_model_head.bin b/loras/opus100/sq/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8415539f27193112869ee4d81511accc6e29c7c4 --- /dev/null +++ b/loras/opus100/sq/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bce5a34ee14feebec881519e50d9880b70636ef0c01593766abb319a9ebd261e +size 342547 diff --git a/loras/opus100/sr/adapter_config.json b/loras/opus100/sr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/sr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/sr/head_config.json b/loras/opus100/sr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/sr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/sr/pytorch_adapter.bin b/loras/opus100/sr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..01b82554779893798adc0633d9c35dc848463bb0 --- /dev/null +++ b/loras/opus100/sr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f798b335f2eb3b1dabd6982aa880733137efac242cf0306ec22c8cb24806d18 +size 1333391 diff --git a/loras/opus100/sr/pytorch_model_head.bin b/loras/opus100/sr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..4d9476b361beed71946499e286649a4c45b9ac29 --- /dev/null +++ b/loras/opus100/sr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1ab2cb4ff5de5f3f117ba4e7ab84dab3ddc6b51741b0b410e0e83b2bf5a700bc +size 342547 diff --git a/loras/opus100/sv/adapter_config.json b/loras/opus100/sv/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/sv/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/sv/head_config.json b/loras/opus100/sv/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/sv/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/sv/pytorch_adapter.bin b/loras/opus100/sv/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..08323094e50c61d72061f041191c96cec5feebdf --- /dev/null +++ b/loras/opus100/sv/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91f1046f80fe4b7a7ede5c544b0efce530ad4ac42f3972e21082f6cd8e91317f +size 1333391 diff --git a/loras/opus100/sv/pytorch_model_head.bin b/loras/opus100/sv/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..236b28b27f09879db76daa84a8b3cf8b2592d49a --- /dev/null +++ b/loras/opus100/sv/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8085199a27000e4999a25aba64c8dea21525f307008422b4ca85ad5eff0d924e +size 342547 diff --git a/loras/opus100/ta/adapter_config.json b/loras/opus100/ta/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ta/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ta/head_config.json b/loras/opus100/ta/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ta/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ta/pytorch_adapter.bin b/loras/opus100/ta/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..dbb05cbf1af6c51a63dae04d5e2e040e9f1e03f8 --- /dev/null +++ b/loras/opus100/ta/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f437e4cdc74fc8fffe91ebc999e246bddb624a8b3b872ffddf78b32b587f6496 +size 1333391 diff --git a/loras/opus100/ta/pytorch_model_head.bin b/loras/opus100/ta/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..1554e9888e64e285935798484a90d08d1cd110ff --- /dev/null +++ b/loras/opus100/ta/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52715fc9050ec0df9de690eb8fb97d92747768b7d2fd54360f2362d0fd4b39e5 +size 342547 diff --git a/loras/opus100/te/adapter_config.json b/loras/opus100/te/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/te/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/te/head_config.json b/loras/opus100/te/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/te/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/te/pytorch_adapter.bin b/loras/opus100/te/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..87ef8210922e870957c69b05596d01ca63805087 --- /dev/null +++ b/loras/opus100/te/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b75dfc12200e445dba7cb4f10b68ef01767fd7c46580c2d72036979009f91df +size 1333391 diff --git a/loras/opus100/te/pytorch_model_head.bin b/loras/opus100/te/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..91a6892972e50fe30d9ac7f11a6e849e749e7eaa --- /dev/null +++ b/loras/opus100/te/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09c97616e0690d312a25d791ddca4d7a83d6807d178e2275349b21a00128d52a +size 342547 diff --git a/loras/opus100/tg/adapter_config.json b/loras/opus100/tg/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/tg/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/tg/head_config.json b/loras/opus100/tg/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/tg/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/tg/pytorch_adapter.bin b/loras/opus100/tg/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..754adbe7f3424cfc8d2f1c2cdd3adaef41b6429c --- /dev/null +++ b/loras/opus100/tg/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c963933a97fa4e6acfb3619e1d2c82706d846f3428743f2c38b6e42c832d67e +size 1333391 diff --git a/loras/opus100/tg/pytorch_model_head.bin b/loras/opus100/tg/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..9c3a717031275b6f75a83c4e7b77b8706f016cb8 --- /dev/null +++ b/loras/opus100/tg/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b027748f2f30d98f128f35bc9b61935fdb17c48a29acab78028a585325c6320 +size 342547 diff --git a/loras/opus100/th/adapter_config.json b/loras/opus100/th/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/th/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/th/head_config.json b/loras/opus100/th/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/th/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/th/pytorch_adapter.bin b/loras/opus100/th/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..eda427eed2cb86f385d22a3ea539f2b09a028014 --- /dev/null +++ b/loras/opus100/th/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:413b041fb84f0bff0d82abd5428b107b6cca09ef937f066fbc09be6725511111 +size 1333391 diff --git a/loras/opus100/th/pytorch_model_head.bin b/loras/opus100/th/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..e0e8f18401e2ef2b1942ffc99107d45b717b40c5 --- /dev/null +++ b/loras/opus100/th/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e20a6b860b44e259f837c834c259abde85f522172b853e3d380208c6e0919587 +size 342547 diff --git a/loras/opus100/tr/adapter_config.json b/loras/opus100/tr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/tr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/tr/head_config.json b/loras/opus100/tr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/tr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/tr/pytorch_adapter.bin b/loras/opus100/tr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..189c4427aa48bb0b672ef0087dd9f9d6dcdd3507 --- /dev/null +++ b/loras/opus100/tr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c96fe1ca152f23c81222de02383684c4e046d67b8b9c80e420c73084176a7b68 +size 1333391 diff --git a/loras/opus100/tr/pytorch_model_head.bin b/loras/opus100/tr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..aa5e78c3b42440174bdaff487ec1a9d63f862108 --- /dev/null +++ b/loras/opus100/tr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9aeb0076aa4952e004209cf680199b51d7cd3d3c085c39e160e8781ea8de3eca +size 342547 diff --git a/loras/opus100/uk/adapter_config.json b/loras/opus100/uk/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/uk/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/uk/head_config.json b/loras/opus100/uk/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/uk/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/uk/pytorch_adapter.bin b/loras/opus100/uk/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..df8b087ea4e75b06165dddc9b34680bc44d40b8e --- /dev/null +++ b/loras/opus100/uk/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a42b435f79a4725d5eabaf7973c253b377e6a0fc41ef35ef68125ace0b6604cb +size 1333391 diff --git a/loras/opus100/uk/pytorch_model_head.bin b/loras/opus100/uk/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..b1207dc16b2875ecd8c3875293f934048e570b76 --- /dev/null +++ b/loras/opus100/uk/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:536af97f777070ae0a891c0f79c3a5cd4348e2c60d758f11d9c4c9afd551be20 +size 342547 diff --git a/loras/opus100/ur/adapter_config.json b/loras/opus100/ur/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/ur/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ur/head_config.json b/loras/opus100/ur/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/ur/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/ur/pytorch_adapter.bin b/loras/opus100/ur/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..e09806cf986611d2ee984376f37eb2f543dd9486 --- /dev/null +++ b/loras/opus100/ur/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59e2e1ee988a65ae7283b146279e1e25a4bf4ef0ca2d51dc7c2d4aef8168fd55 +size 1333391 diff --git a/loras/opus100/ur/pytorch_model_head.bin b/loras/opus100/ur/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..d2cf40063d98787eeb4d7c36b2fd45a878246ffe --- /dev/null +++ b/loras/opus100/ur/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a92658680a9aba0c03e217cef97a2e747cb933e587e1e850072ba720533b8069 +size 342547 diff --git a/loras/opus100/uz/adapter_config.json b/loras/opus100/uz/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/uz/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/uz/head_config.json b/loras/opus100/uz/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/uz/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/uz/pytorch_adapter.bin b/loras/opus100/uz/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ec26d596e73912ee4ab0d8d2a340c18ab80ef7d0 --- /dev/null +++ b/loras/opus100/uz/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1ee8488ba0e6018b781c1d2224e743c4dae3242d7547ba266f2fc01d8849be81 +size 1333391 diff --git a/loras/opus100/uz/pytorch_model_head.bin b/loras/opus100/uz/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..e6753cf0df387851b02af833a1b57dd7948d1d38 --- /dev/null +++ b/loras/opus100/uz/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0399ada67fc4f096ca9676519f8d5273a854a7f72aec7b0cf239a5a97f2c574 +size 342547 diff --git a/loras/opus100/vi/adapter_config.json b/loras/opus100/vi/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/vi/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/vi/head_config.json b/loras/opus100/vi/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/vi/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/vi/pytorch_adapter.bin b/loras/opus100/vi/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..00f1d99f9b67e96403cf4b2be6623f97a2e18317 --- /dev/null +++ b/loras/opus100/vi/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8079ee87526da08e22b1a86a2d298d543736d8135a75e442887e7c91ac3807fb +size 1333391 diff --git a/loras/opus100/vi/pytorch_model_head.bin b/loras/opus100/vi/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..12013130c212e9d3544817867f2f6c911e795c4a --- /dev/null +++ b/loras/opus100/vi/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6c11a614dc3fe8634df6b72111b360949a769e83bdbed8c0d8841a985f3ec2a +size 342547 diff --git a/loras/opus100/xh/adapter_config.json b/loras/opus100/xh/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/xh/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/xh/head_config.json b/loras/opus100/xh/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/xh/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/xh/pytorch_adapter.bin b/loras/opus100/xh/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..b0da5257e5fcbbf8d84494fd9b60f18a332bfdab --- /dev/null +++ b/loras/opus100/xh/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0fc6c7031ce2c1e2a6207b5b71873fb824c9a2b048045b25bfc4ef1ec2cc76e3 +size 1333391 diff --git a/loras/opus100/xh/pytorch_model_head.bin b/loras/opus100/xh/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..75b45036996b7481c0274bd984efda4a1dfd15fa --- /dev/null +++ b/loras/opus100/xh/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44d28acb5635788d800625e26bea4037afb1bf2c44525924bc79f551d21bbb1d +size 342547 diff --git a/loras/opus100/yi/adapter_config.json b/loras/opus100/yi/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/yi/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/yi/head_config.json b/loras/opus100/yi/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/yi/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/yi/pytorch_adapter.bin b/loras/opus100/yi/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..7d669498b0d40e9ef8be549469e839c1e217f5dd --- /dev/null +++ b/loras/opus100/yi/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3372403e1f98da41dcff8b7f0eb01fadaf6e8b4f60cc1abf4e4ccaa8a5bbfc8b +size 1333391 diff --git a/loras/opus100/yi/pytorch_model_head.bin b/loras/opus100/yi/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..0da36f7fcc5433f88d48f10b855b80ced4616902 --- /dev/null +++ b/loras/opus100/yi/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebbb185c3c1d1187df97d62bad637bf3b592bdcfe7e63eda79a629c003fba9d8 +size 342547 diff --git a/loras/opus100/zh/adapter_config.json b/loras/opus100/zh/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/zh/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/zh/head_config.json b/loras/opus100/zh/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/zh/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/zh/pytorch_adapter.bin b/loras/opus100/zh/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..9180fd89230e12c3c82c168392d4cb1f859b30a8 --- /dev/null +++ b/loras/opus100/zh/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:367f78fa1fc1f0d2c31ae4c42e10712ade48481e7607fae5c2204d91196cb8cb +size 1333391 diff --git a/loras/opus100/zh/pytorch_model_head.bin b/loras/opus100/zh/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..f31ed34c04de14c4cdb484332beb9ed2f7ed032b --- /dev/null +++ b/loras/opus100/zh/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d57349ea2010e3be0dbe4c180300931718813b95268b023c6dfc275226ef35c +size 342547 diff --git a/loras/opus100/zu/adapter_config.json b/loras/opus100/zu/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/opus100/zu/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/zu/head_config.json b/loras/opus100/zu/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/opus100/zu/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/opus100/zu/pytorch_adapter.bin b/loras/opus100/zu/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ed1830e11cd1ffa7b3ae5084ae276efed0181031 --- /dev/null +++ b/loras/opus100/zu/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f96d31b48e565f1b0ff402ab521b6b01ef65b89bce8a09faa412c107595218c8 +size 1333391 diff --git a/loras/opus100/zu/pytorch_model_head.bin b/loras/opus100/zu/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..cff902ad5f88b6e723f8ecdc5988a377e346a852 --- /dev/null +++ b/loras/opus100/zu/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:181e48d2c16d419d0bce0ee966bf2a4ab3dad7737ed2ada8ba361bec86737831 +size 342547 diff --git a/loras/ted2020-corrupted/af/adapter_config.json b/loras/ted2020-corrupted/af/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/af/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/af/head_config.json b/loras/ted2020-corrupted/af/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/af/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/af/pytorch_adapter.bin b/loras/ted2020-corrupted/af/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..301891d2655f994455831bad9a66953e775366a0 --- /dev/null +++ b/loras/ted2020-corrupted/af/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b3b67cacca2dee46f79a96841e4674e89f84e9545173ccdc025ec526e71ce09 +size 1333391 diff --git a/loras/ted2020-corrupted/af/pytorch_model_head.bin b/loras/ted2020-corrupted/af/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..fde53cc0c43eb9014134d08c933fb0857e33ba37 --- /dev/null +++ b/loras/ted2020-corrupted/af/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a931f919e5aeab6e3a613d9f927bab35518a96033f0ac5f3ec23c01a7197e950 +size 342547 diff --git a/loras/ted2020-corrupted/am/adapter_config.json b/loras/ted2020-corrupted/am/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/am/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/am/head_config.json b/loras/ted2020-corrupted/am/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/am/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/am/pytorch_adapter.bin b/loras/ted2020-corrupted/am/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..bd1934420acde46fdb91b86ded61619cdaaf6478 --- /dev/null +++ b/loras/ted2020-corrupted/am/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4ad757107d946b07a6ff67704b11e4683719757d0a76f2eccc5b5dc0d702ed0 +size 1333391 diff --git a/loras/ted2020-corrupted/am/pytorch_model_head.bin b/loras/ted2020-corrupted/am/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..203d34c58c1d9facdddf15256b078770d6c246cd --- /dev/null +++ b/loras/ted2020-corrupted/am/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff1060048474f43dfcb29562b75ee161929e8cd903e6c5e625bb36f9051b3cb1 +size 342547 diff --git a/loras/ted2020-corrupted/ar/adapter_config.json b/loras/ted2020-corrupted/ar/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ar/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ar/head_config.json b/loras/ted2020-corrupted/ar/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ar/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ar/pytorch_adapter.bin b/loras/ted2020-corrupted/ar/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..f3e0544cb0c7d4b1dcd55f1d7581473aca7a42c2 --- /dev/null +++ b/loras/ted2020-corrupted/ar/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96317e217771593c4631a7be454820ed51d4a45fb170b946e7c1db37ad7d36ce +size 1333391 diff --git a/loras/ted2020-corrupted/ar/pytorch_model_head.bin b/loras/ted2020-corrupted/ar/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..e9915f9c6f881e9b7ee00cddcda0757e3bc9272b --- /dev/null +++ b/loras/ted2020-corrupted/ar/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59bc361abbd7efcab48b713da73dbb30c82374b01cdb8bca0d43e218842e0ef2 +size 342547 diff --git a/loras/ted2020-corrupted/az/adapter_config.json b/loras/ted2020-corrupted/az/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/az/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/az/head_config.json b/loras/ted2020-corrupted/az/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/az/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/az/pytorch_adapter.bin b/loras/ted2020-corrupted/az/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..c824e865d2650af2cb63f9cedc6186e338d257c3 --- /dev/null +++ b/loras/ted2020-corrupted/az/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e25464db0aefe4e52a2804fdee50fae794370cda4240ee3a1e076d8a605c3f68 +size 1333391 diff --git a/loras/ted2020-corrupted/az/pytorch_model_head.bin b/loras/ted2020-corrupted/az/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..3ab0581f49feac1d52577d3610a7615ef0f5a8e4 --- /dev/null +++ b/loras/ted2020-corrupted/az/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7376a06ed9bedfd3d09b5ce553c510f727f496f61db6d3ebfad2d2580022c271 +size 342547 diff --git a/loras/ted2020-corrupted/be/adapter_config.json b/loras/ted2020-corrupted/be/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/be/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/be/head_config.json b/loras/ted2020-corrupted/be/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/be/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/be/pytorch_adapter.bin b/loras/ted2020-corrupted/be/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ded9f4adfcdaa4d6f46c0546520547519e78f6ca --- /dev/null +++ b/loras/ted2020-corrupted/be/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef6c5a07710fa1d509e9f9afa2a0d9a377f240fce8a1a276e111b80c80876502 +size 1333391 diff --git a/loras/ted2020-corrupted/be/pytorch_model_head.bin b/loras/ted2020-corrupted/be/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..43067c1ac7cdab0a264e496113d6d760c1341d65 --- /dev/null +++ b/loras/ted2020-corrupted/be/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:623589b18368af6796791545af567792b39f0815deb44fdc8245bdd5f682a474 +size 342547 diff --git a/loras/ted2020-corrupted/bg/adapter_config.json b/loras/ted2020-corrupted/bg/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/bg/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/bg/head_config.json b/loras/ted2020-corrupted/bg/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/bg/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/bg/pytorch_adapter.bin b/loras/ted2020-corrupted/bg/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..b0c368c1539e985b5bd653175633fb153e99eae8 --- /dev/null +++ b/loras/ted2020-corrupted/bg/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c17de975380150ae4466c98b86662e884f0ac27d1a6164b7ba153d688d69430 +size 1333391 diff --git a/loras/ted2020-corrupted/bg/pytorch_model_head.bin b/loras/ted2020-corrupted/bg/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..135f2b56e5a8d5f060552418230b3bfcd197c768 --- /dev/null +++ b/loras/ted2020-corrupted/bg/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf68409e245422b7eb6b6cadb4f23e99944623678d0b5c227dff81a7ad93554c +size 342547 diff --git a/loras/ted2020-corrupted/bn/adapter_config.json b/loras/ted2020-corrupted/bn/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/bn/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/bn/head_config.json b/loras/ted2020-corrupted/bn/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/bn/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/bn/pytorch_adapter.bin b/loras/ted2020-corrupted/bn/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..15f8ec86add72aa24b9cae0d1762ee8913ad15be --- /dev/null +++ b/loras/ted2020-corrupted/bn/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7625ce1d970113d0c6a0be90fca919a8b10ebaed04395dab486f769e1cf9324e +size 1333391 diff --git a/loras/ted2020-corrupted/bn/pytorch_model_head.bin b/loras/ted2020-corrupted/bn/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..c40e00f22800998f1f38f61c179a283c887c0526 --- /dev/null +++ b/loras/ted2020-corrupted/bn/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:798c941fd10edccf3f7677e44d081fa6fbd536787aa22c7b1e496dff7231a8f8 +size 342547 diff --git a/loras/ted2020-corrupted/ca/adapter_config.json b/loras/ted2020-corrupted/ca/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ca/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ca/head_config.json b/loras/ted2020-corrupted/ca/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ca/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ca/pytorch_adapter.bin b/loras/ted2020-corrupted/ca/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..67dc906c526675a8690bab55c3a431382e35fbf5 --- /dev/null +++ b/loras/ted2020-corrupted/ca/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ad43c22cd5b3f81156b0e5dcbc2ea286cc0a987bb66edc5f3bfcdb12a82201e +size 1333391 diff --git a/loras/ted2020-corrupted/ca/pytorch_model_head.bin b/loras/ted2020-corrupted/ca/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..982e3c16c0cb6cdb5a687a8a1b0fc459ac95c53f --- /dev/null +++ b/loras/ted2020-corrupted/ca/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4dc95a87e1d9b65a33549eeccaa931994538411543ce9957b1e06327353c15b0 +size 342547 diff --git a/loras/ted2020-corrupted/ceb/adapter_config.json b/loras/ted2020-corrupted/ceb/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ceb/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ceb/head_config.json b/loras/ted2020-corrupted/ceb/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ceb/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ceb/pytorch_adapter.bin b/loras/ted2020-corrupted/ceb/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d989e4552f7e34fa3d4c0c38f0e6059fb71dc26b --- /dev/null +++ b/loras/ted2020-corrupted/ceb/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6d966446eb8d081a408a4b1c95ba9b085cda1fa01580d7bc8f816e2887f4952 +size 1333391 diff --git a/loras/ted2020-corrupted/ceb/pytorch_model_head.bin b/loras/ted2020-corrupted/ceb/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..7a48d8b485e0f3e378a20726437242e4839b8357 --- /dev/null +++ b/loras/ted2020-corrupted/ceb/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:75488093b0d8a55c0622674506ce44bf5ad377c9efb5302cd984f3ff6a485cff +size 342547 diff --git a/loras/ted2020-corrupted/cs/adapter_config.json b/loras/ted2020-corrupted/cs/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/cs/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/cs/head_config.json b/loras/ted2020-corrupted/cs/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/cs/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/cs/pytorch_adapter.bin b/loras/ted2020-corrupted/cs/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..b6aaf2885ec463f9e4a92ea076740a30fd2aa938 --- /dev/null +++ b/loras/ted2020-corrupted/cs/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b6238c8a6cd68616a035ae979133d3110b36c6f58d61bd5bf5120e548ac9e4a +size 1333391 diff --git a/loras/ted2020-corrupted/cs/pytorch_model_head.bin b/loras/ted2020-corrupted/cs/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..814f56d465060f2a052b65e9c518317d5be13053 --- /dev/null +++ b/loras/ted2020-corrupted/cs/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f77b84c6e727b7791d3ba5456d20008d2e63a9570f068c48c87093d873382ba +size 342547 diff --git a/loras/ted2020-corrupted/da/adapter_config.json b/loras/ted2020-corrupted/da/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/da/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/da/head_config.json b/loras/ted2020-corrupted/da/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/da/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/da/pytorch_adapter.bin b/loras/ted2020-corrupted/da/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..c9a9fa1f14589e0405e8ccab1e1dffcaad8f99f4 --- /dev/null +++ b/loras/ted2020-corrupted/da/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fa4abf7a354799fa22300b6121bf878354d61aafa740c06ded8a7e8b25e2bff +size 1333391 diff --git a/loras/ted2020-corrupted/da/pytorch_model_head.bin b/loras/ted2020-corrupted/da/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..1d2745f948cbf00b9c51e428cb3055a30d0b5f80 --- /dev/null +++ b/loras/ted2020-corrupted/da/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d2ae27c7e2f8b9b090ff3fe050959e15d41d4f3f43858350ef008e004e33499 +size 342547 diff --git a/loras/ted2020-corrupted/de/adapter_config.json b/loras/ted2020-corrupted/de/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/de/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/de/head_config.json b/loras/ted2020-corrupted/de/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/de/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/de/pytorch_adapter.bin b/loras/ted2020-corrupted/de/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..a75170028f34d5d158086a76621945ff14dfb5e3 --- /dev/null +++ b/loras/ted2020-corrupted/de/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cce6abf68d75f04b4d49c2374890fe112c000a8f6642f9ec319098c929bd0056 +size 1333391 diff --git a/loras/ted2020-corrupted/de/pytorch_model_head.bin b/loras/ted2020-corrupted/de/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..0ba4bcc4e267f626d037ac75fb0c6c881ddebb93 --- /dev/null +++ b/loras/ted2020-corrupted/de/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45525c719e836b83684c0f52325653f47b37b45f7c3a255cd1f2d5651b1eab95 +size 342547 diff --git a/loras/ted2020-corrupted/el/adapter_config.json b/loras/ted2020-corrupted/el/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/el/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/el/head_config.json b/loras/ted2020-corrupted/el/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/el/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/el/pytorch_adapter.bin b/loras/ted2020-corrupted/el/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d4a3756836277517a9e98c2d50a53ad4088eeff1 --- /dev/null +++ b/loras/ted2020-corrupted/el/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4863c433a51e1804f46bf6f79b4bc1b102f76f93576be4583d80b126c97ecfee +size 1333391 diff --git a/loras/ted2020-corrupted/el/pytorch_model_head.bin b/loras/ted2020-corrupted/el/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..1d9eae39322a365d275bd12fd4de0a4b173c7931 --- /dev/null +++ b/loras/ted2020-corrupted/el/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1ace2e1b478341867717dcc6a6fe9550da7cd376586b07c7e4f3f18cd184d826 +size 342547 diff --git a/loras/ted2020-corrupted/en/adapter_config.json b/loras/ted2020-corrupted/en/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/en/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/en/head_config.json b/loras/ted2020-corrupted/en/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/en/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/en/pytorch_adapter.bin b/loras/ted2020-corrupted/en/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d6194ac4fb7f414bbbb4973f82cb57f94aab1909 --- /dev/null +++ b/loras/ted2020-corrupted/en/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:656215a144853d8d873495b26c908241531543f6692d27b4a6e224a930dc686e +size 1333391 diff --git a/loras/ted2020-corrupted/en/pytorch_model_head.bin b/loras/ted2020-corrupted/en/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..3da78527227479f1fcc47a66d5624dee1946a65d --- /dev/null +++ b/loras/ted2020-corrupted/en/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55cb875315ea9ff37e9a1e115243225b090892107126870069018ef42fe5476e +size 342547 diff --git a/loras/ted2020-corrupted/eo/adapter_config.json b/loras/ted2020-corrupted/eo/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/eo/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/eo/head_config.json b/loras/ted2020-corrupted/eo/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/eo/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/eo/pytorch_adapter.bin b/loras/ted2020-corrupted/eo/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ec9c68ab81137726f8815c07b6b60b2b61d76c69 --- /dev/null +++ b/loras/ted2020-corrupted/eo/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:360a6ccc3ab39193c2e71354d548f5bbe4e8422a0db56f3042c3fd9373739026 +size 1333391 diff --git a/loras/ted2020-corrupted/eo/pytorch_model_head.bin b/loras/ted2020-corrupted/eo/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..fab393e0f419571069fa69cea86d0eea2b912058 --- /dev/null +++ b/loras/ted2020-corrupted/eo/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4da78dbba0af9c8a8cc985a83b9f7a602700d43e1501c0147f3de9801aefba79 +size 342547 diff --git a/loras/ted2020-corrupted/es/adapter_config.json b/loras/ted2020-corrupted/es/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/es/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/es/head_config.json b/loras/ted2020-corrupted/es/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/es/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/es/pytorch_adapter.bin b/loras/ted2020-corrupted/es/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..2e9372771728e0ec45cc0263197275661456d806 --- /dev/null +++ b/loras/ted2020-corrupted/es/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab8cdee0f505f5cf184486fb5b7cb7a6758c29e7a646735fcbffed091ca8f808 +size 1333391 diff --git a/loras/ted2020-corrupted/es/pytorch_model_head.bin b/loras/ted2020-corrupted/es/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8cd4c1803c09405df028217530f2d09e823d31b1 --- /dev/null +++ b/loras/ted2020-corrupted/es/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2405b14be3789895827722a90b6fed242351b89a61b8e8abeab29d5c9de782f1 +size 342547 diff --git a/loras/ted2020-corrupted/et/adapter_config.json b/loras/ted2020-corrupted/et/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/et/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/et/head_config.json b/loras/ted2020-corrupted/et/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/et/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/et/pytorch_adapter.bin b/loras/ted2020-corrupted/et/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..5ee8def1ea313e0d88835945a4c6922b97a4972e --- /dev/null +++ b/loras/ted2020-corrupted/et/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14f3b0c25df027fb0ed0fa7f11831f6fd0f5874893134a40aa45001fe357fc2b +size 1333391 diff --git a/loras/ted2020-corrupted/et/pytorch_model_head.bin b/loras/ted2020-corrupted/et/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..74d48c54df40e133b007be1debeec2a38347880a --- /dev/null +++ b/loras/ted2020-corrupted/et/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a39bb16704ba68f418c7604b43d46cf40fdf03b172051057fb80fe9daadb8076 +size 342547 diff --git a/loras/ted2020-corrupted/eu/adapter_config.json b/loras/ted2020-corrupted/eu/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/eu/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/eu/head_config.json b/loras/ted2020-corrupted/eu/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/eu/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/eu/pytorch_adapter.bin b/loras/ted2020-corrupted/eu/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1b230a3c55591df0604d438acfa46ee56a74f44f --- /dev/null +++ b/loras/ted2020-corrupted/eu/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8db8c5433a3d8702e2134e525cbe80a4b02c4563990c2604b5dc1fb23620e1be +size 1333391 diff --git a/loras/ted2020-corrupted/eu/pytorch_model_head.bin b/loras/ted2020-corrupted/eu/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..4e0fc65848a2a567f2b3cd043c1be0b8da1f491a --- /dev/null +++ b/loras/ted2020-corrupted/eu/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91c4ed4449a44f1d7f8e02a96161111dc7e38d5bfb824cb8afd2d7b1ec90d621 +size 342547 diff --git a/loras/ted2020-corrupted/fa/adapter_config.json b/loras/ted2020-corrupted/fa/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/fa/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/fa/head_config.json b/loras/ted2020-corrupted/fa/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/fa/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/fa/pytorch_adapter.bin b/loras/ted2020-corrupted/fa/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d1d269b56a44d84a170879e8c47f2468f85e5d16 --- /dev/null +++ b/loras/ted2020-corrupted/fa/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cddf298afe08f3e4a56e26afaa49dd46eef871c339dd4cbba95128a64498241e +size 1333391 diff --git a/loras/ted2020-corrupted/fa/pytorch_model_head.bin b/loras/ted2020-corrupted/fa/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..6c03e2123708e48ac3b542b45b34d1337562375e --- /dev/null +++ b/loras/ted2020-corrupted/fa/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6a0271a7c42d2649be5baefb3a3cdd7943761a5e36e033a5023f78e01848fb0 +size 342547 diff --git a/loras/ted2020-corrupted/fi/adapter_config.json b/loras/ted2020-corrupted/fi/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/fi/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/fi/head_config.json b/loras/ted2020-corrupted/fi/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/fi/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/fi/pytorch_adapter.bin b/loras/ted2020-corrupted/fi/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..a63d61d089131ed2bce0ed182ec03ca6a61946ec --- /dev/null +++ b/loras/ted2020-corrupted/fi/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2217c1e218a5bb79145cdc6318022965da073a821611f0712972f15629b2139f +size 1333391 diff --git a/loras/ted2020-corrupted/fi/pytorch_model_head.bin b/loras/ted2020-corrupted/fi/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..cbd72b76a62627d291c2934bd5a4de55b0d216a9 --- /dev/null +++ b/loras/ted2020-corrupted/fi/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2bed6aea523fe7383812b31d7335eee046b2f819d491298f94d208d5d8d23bb4 +size 342547 diff --git a/loras/ted2020-corrupted/fr/adapter_config.json b/loras/ted2020-corrupted/fr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/fr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/fr/head_config.json b/loras/ted2020-corrupted/fr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/fr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/fr/pytorch_adapter.bin b/loras/ted2020-corrupted/fr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..526695fd5d13cc2062d1d2f90898975476609967 --- /dev/null +++ b/loras/ted2020-corrupted/fr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3667a752c0ec4362a7703869013c8c916ecb85120483ca49134ee10b051f7e85 +size 1333391 diff --git a/loras/ted2020-corrupted/fr/pytorch_model_head.bin b/loras/ted2020-corrupted/fr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..738ecc8a55aff7458e5c50d0bf2ddc7d017c7f58 --- /dev/null +++ b/loras/ted2020-corrupted/fr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b507c65f2c75cfd19e79a45b4f86f736a3b01a946e21894eca652c80b0c7a86 +size 342547 diff --git a/loras/ted2020-corrupted/ga/adapter_config.json b/loras/ted2020-corrupted/ga/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ga/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ga/head_config.json b/loras/ted2020-corrupted/ga/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ga/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ga/pytorch_adapter.bin b/loras/ted2020-corrupted/ga/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d7aa7f9d8ec2cd22ebe002f0732c6146bbf27705 --- /dev/null +++ b/loras/ted2020-corrupted/ga/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f70eed0c6eff3b3861e98c1dcbc884d1b53b2ff546cbeb283bbb5090c110ef23 +size 1333391 diff --git a/loras/ted2020-corrupted/ga/pytorch_model_head.bin b/loras/ted2020-corrupted/ga/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..4da3208186c02b73ce962b439978d6106d200ada --- /dev/null +++ b/loras/ted2020-corrupted/ga/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79bb51d29165f89fd4f53baca0b43485d63e76911da70db16b486bc8b4b0c508 +size 342547 diff --git a/loras/ted2020-corrupted/gl/adapter_config.json b/loras/ted2020-corrupted/gl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/gl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/gl/head_config.json b/loras/ted2020-corrupted/gl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/gl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/gl/pytorch_adapter.bin b/loras/ted2020-corrupted/gl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..f20c8b8217e8d365f4c1b382b010af92af8e1c91 --- /dev/null +++ b/loras/ted2020-corrupted/gl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08476e5cc86c6fd29611542d1d017a9f02d0029af2cdcd3f31c9ee9dd6f2d8cb +size 1333391 diff --git a/loras/ted2020-corrupted/gl/pytorch_model_head.bin b/loras/ted2020-corrupted/gl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..c2e9389cb0ed44003a691e6eedea45d543b073cd --- /dev/null +++ b/loras/ted2020-corrupted/gl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ddaac3dcac35367eaeebedcac15bcef6f5a20714a656b24c72f313c28518e990 +size 342547 diff --git a/loras/ted2020-corrupted/gu/adapter_config.json b/loras/ted2020-corrupted/gu/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/gu/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/gu/head_config.json b/loras/ted2020-corrupted/gu/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/gu/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/gu/pytorch_adapter.bin b/loras/ted2020-corrupted/gu/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d7eb9a6e5ddfce4c49ab61e72cd82adc6051afb6 --- /dev/null +++ b/loras/ted2020-corrupted/gu/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d512cf608d056f5b1643c9829bd2c8d13c912d53a490ffeab0a7c433f2580425 +size 1333391 diff --git a/loras/ted2020-corrupted/gu/pytorch_model_head.bin b/loras/ted2020-corrupted/gu/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..4cfc96acab99c7f3bc66428b4e66fa57e9b423cd --- /dev/null +++ b/loras/ted2020-corrupted/gu/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:411603b4de7a07c944c0e88bc5f6d164d489b2428ff7ba6271a43bbfc9af4101 +size 342547 diff --git a/loras/ted2020-corrupted/ha/adapter_config.json b/loras/ted2020-corrupted/ha/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ha/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ha/head_config.json b/loras/ted2020-corrupted/ha/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ha/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ha/pytorch_adapter.bin b/loras/ted2020-corrupted/ha/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1b2eacea2dccedb5bb3c774a216d8273050b451b --- /dev/null +++ b/loras/ted2020-corrupted/ha/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:616e779a5e92735902f9dd69f934e7a6118863af51fbddd0d4ed589d02d44066 +size 1333391 diff --git a/loras/ted2020-corrupted/ha/pytorch_model_head.bin b/loras/ted2020-corrupted/ha/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..88a7dcb50d3508e8d7973aa27bf1cc96094bfde4 --- /dev/null +++ b/loras/ted2020-corrupted/ha/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d5acec114d03a6135251d4c95384b99bff1d77a68f7bff16d453f5f86c7804b3 +size 342547 diff --git a/loras/ted2020-corrupted/he/adapter_config.json b/loras/ted2020-corrupted/he/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/he/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/he/head_config.json b/loras/ted2020-corrupted/he/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/he/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/he/pytorch_adapter.bin b/loras/ted2020-corrupted/he/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..f99da6c409aac7995b21d26c079321604c4a871c --- /dev/null +++ b/loras/ted2020-corrupted/he/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:027117f42f677b5283050e6f15177066c75dfd41920de7b4b693ea544769dea7 +size 1333391 diff --git a/loras/ted2020-corrupted/he/pytorch_model_head.bin b/loras/ted2020-corrupted/he/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..6a4460ffc9b9c20791a5544e43d9a248c8167958 --- /dev/null +++ b/loras/ted2020-corrupted/he/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27ee0fbc6bf5f2e7798fa886abae2d16516160c57cf0799cc897bc475ed27170 +size 342547 diff --git a/loras/ted2020-corrupted/hi/adapter_config.json b/loras/ted2020-corrupted/hi/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/hi/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/hi/head_config.json b/loras/ted2020-corrupted/hi/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/hi/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/hi/pytorch_adapter.bin b/loras/ted2020-corrupted/hi/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..44c1b06b737b02d5628cb2a4606bdc9106e9f090 --- /dev/null +++ b/loras/ted2020-corrupted/hi/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5db4daacf764be35b9a32c6de0cc375ff5f8397104bfdaaad021d72661665daf +size 1333391 diff --git a/loras/ted2020-corrupted/hi/pytorch_model_head.bin b/loras/ted2020-corrupted/hi/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..5256ad3e0a680b16cb527a01247c884f81d0fe1f --- /dev/null +++ b/loras/ted2020-corrupted/hi/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e42f39313c406932df46d6ff9d0e1dcadcc5cc314f8c1f54ee14fdeb5c6960c +size 342547 diff --git a/loras/ted2020-corrupted/hu/adapter_config.json b/loras/ted2020-corrupted/hu/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/hu/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/hu/head_config.json b/loras/ted2020-corrupted/hu/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/hu/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/hu/pytorch_adapter.bin b/loras/ted2020-corrupted/hu/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ae09c5e747602450a1aa449eac6967ede0eed645 --- /dev/null +++ b/loras/ted2020-corrupted/hu/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8939ed00762199658c52d6737da8aeac123deeef21d0425919c118ec463c47c3 +size 1333391 diff --git a/loras/ted2020-corrupted/hu/pytorch_model_head.bin b/loras/ted2020-corrupted/hu/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..c7c9f19d6e3e5b20bd1d255a44ed5c1911860173 --- /dev/null +++ b/loras/ted2020-corrupted/hu/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f791ae7c8bf62fd0044e2cd1d4366d72813be2de8c670de0f9324a1019930eb +size 342547 diff --git a/loras/ted2020-corrupted/hy/adapter_config.json b/loras/ted2020-corrupted/hy/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/hy/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/hy/head_config.json b/loras/ted2020-corrupted/hy/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/hy/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/hy/pytorch_adapter.bin b/loras/ted2020-corrupted/hy/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..a684390a23c331f2fc3180d769a243f6fd367e84 --- /dev/null +++ b/loras/ted2020-corrupted/hy/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0eebabdec016442364b60fae9e128dee2f01ad7fc83a1d404e8396d178906bdf +size 1333391 diff --git a/loras/ted2020-corrupted/hy/pytorch_model_head.bin b/loras/ted2020-corrupted/hy/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..2fe5444859212eb7d74a2a8a70c8aaa88c82a5b6 --- /dev/null +++ b/loras/ted2020-corrupted/hy/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9fce3b6f4decdd025fc38e43afe04105503b23923370ee4100b7be4166ca2a8b +size 342547 diff --git a/loras/ted2020-corrupted/id/adapter_config.json b/loras/ted2020-corrupted/id/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/id/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/id/head_config.json b/loras/ted2020-corrupted/id/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/id/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/id/pytorch_adapter.bin b/loras/ted2020-corrupted/id/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..42b1f81c8515847d0d3a5fc73fc74ee3b912389f --- /dev/null +++ b/loras/ted2020-corrupted/id/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5845715581f1c262b277be22ba42a60728f45e10a81f8d062f6c6a778566d7d +size 1333391 diff --git a/loras/ted2020-corrupted/id/pytorch_model_head.bin b/loras/ted2020-corrupted/id/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..88e56c0288c1b545103e4ca77326fb9a01fe51c1 --- /dev/null +++ b/loras/ted2020-corrupted/id/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0a0ffbb7de5cdd9486ca23a874adabbddb10c7a81742a6864721dccc09aeb9a +size 342547 diff --git a/loras/ted2020-corrupted/ig/adapter_config.json b/loras/ted2020-corrupted/ig/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ig/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ig/head_config.json b/loras/ted2020-corrupted/ig/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ig/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ig/pytorch_adapter.bin b/loras/ted2020-corrupted/ig/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d31e6d4c39ee03baace4c7c2d2ba28365849acf9 --- /dev/null +++ b/loras/ted2020-corrupted/ig/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:704d508a4be851ce5bd123dd6316cd61f120207f3429064c657c4ebb5b6c82a5 +size 1333391 diff --git a/loras/ted2020-corrupted/ig/pytorch_model_head.bin b/loras/ted2020-corrupted/ig/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..ebacd05810e698ed9d4e78fcb72ca501b88429ae --- /dev/null +++ b/loras/ted2020-corrupted/ig/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef8d7f4c17ee6a7fb04630c3c8841996bded377927e8076e288c41783963a902 +size 342547 diff --git a/loras/ted2020-corrupted/is/adapter_config.json b/loras/ted2020-corrupted/is/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/is/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/is/head_config.json b/loras/ted2020-corrupted/is/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/is/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/is/pytorch_adapter.bin b/loras/ted2020-corrupted/is/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..eb96063f590d8dfb87d8ea44bda1cc43557964f8 --- /dev/null +++ b/loras/ted2020-corrupted/is/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71cd8059f8f9c80af698e9897aefae6c72bb7a5a73c99974be5849f63c315d54 +size 1333391 diff --git a/loras/ted2020-corrupted/is/pytorch_model_head.bin b/loras/ted2020-corrupted/is/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..dc019e7d1ecc2535a21cff60f2a7e23d235c2403 --- /dev/null +++ b/loras/ted2020-corrupted/is/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92657e5de50f7a09630a9b690d65cb655dbb2f583eb8ea4a7b2faf628923a36c +size 342547 diff --git a/loras/ted2020-corrupted/it/adapter_config.json b/loras/ted2020-corrupted/it/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/it/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/it/head_config.json b/loras/ted2020-corrupted/it/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/it/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/it/pytorch_adapter.bin b/loras/ted2020-corrupted/it/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..9421082962874d605b11f9afc5ab6a1c21c5575d --- /dev/null +++ b/loras/ted2020-corrupted/it/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f8d2116e57594bcd1ace414e925060d264058fb7f66eda3dd8edded733dd495 +size 1333391 diff --git a/loras/ted2020-corrupted/it/pytorch_model_head.bin b/loras/ted2020-corrupted/it/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..89bfea0e7536ac0bc5039039b67444b131425d79 --- /dev/null +++ b/loras/ted2020-corrupted/it/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc7adb06e8b2a6ce3cb7fe0b712d7fbe2ce7a41e9f39f504f2ab34462ad3f4e7 +size 342547 diff --git a/loras/ted2020-corrupted/ja/adapter_config.json b/loras/ted2020-corrupted/ja/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ja/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ja/head_config.json b/loras/ted2020-corrupted/ja/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ja/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ja/pytorch_adapter.bin b/loras/ted2020-corrupted/ja/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..64d54a7ad64411d38ea264b3334d696d0f2f114c --- /dev/null +++ b/loras/ted2020-corrupted/ja/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60d22bfc62ecf8300eb7c353424bf3340acd20bce59e7f65ee597642083e1772 +size 1333391 diff --git a/loras/ted2020-corrupted/ja/pytorch_model_head.bin b/loras/ted2020-corrupted/ja/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..7e2de070ac469e16dec1da8b0d808062c8d9b73b --- /dev/null +++ b/loras/ted2020-corrupted/ja/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c851fc9341f7e5bb6ad008b1018657c727c98ee5a67a413c90d5c87190c7f71 +size 342547 diff --git a/loras/ted2020-corrupted/ka/adapter_config.json b/loras/ted2020-corrupted/ka/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ka/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ka/head_config.json b/loras/ted2020-corrupted/ka/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ka/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ka/pytorch_adapter.bin b/loras/ted2020-corrupted/ka/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..34e77664b3b5a3a8e59b0936c95a45043753ace0 --- /dev/null +++ b/loras/ted2020-corrupted/ka/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e335aa6352e63c13ce251cd2ed8e0d07f823c6448d73c0a92a5ef19005f56272 +size 1333391 diff --git a/loras/ted2020-corrupted/ka/pytorch_model_head.bin b/loras/ted2020-corrupted/ka/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..3d1c8962a0fcfc6d822056b5571a22420f4b2bd0 --- /dev/null +++ b/loras/ted2020-corrupted/ka/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f40c348a742d497490af3b866dcfd21ab054a3bb7c75a5d5f55b5e8b299a604 +size 342547 diff --git a/loras/ted2020-corrupted/kk/adapter_config.json b/loras/ted2020-corrupted/kk/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/kk/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/kk/head_config.json b/loras/ted2020-corrupted/kk/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/kk/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/kk/pytorch_adapter.bin b/loras/ted2020-corrupted/kk/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..295ea1c2a43ffa65600a25373a0d39dd19e71ab9 --- /dev/null +++ b/loras/ted2020-corrupted/kk/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3456f866045f86b8df8f8e7a48c0fedb4b49fd5f6ed80b52cb28fed8ba17ece +size 1333391 diff --git a/loras/ted2020-corrupted/kk/pytorch_model_head.bin b/loras/ted2020-corrupted/kk/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..77eca25b7599c23eaeaf0ce1c60cd53c36914321 --- /dev/null +++ b/loras/ted2020-corrupted/kk/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93d41d2491e6ca25d19624c3cbb5725d3c5fa740d46e71d5fc0a1a5bdc5ac910 +size 342547 diff --git a/loras/ted2020-corrupted/km/adapter_config.json b/loras/ted2020-corrupted/km/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/km/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/km/head_config.json b/loras/ted2020-corrupted/km/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/km/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/km/pytorch_adapter.bin b/loras/ted2020-corrupted/km/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..07f23f96a6520c00ae06159942089f45c2b48e50 --- /dev/null +++ b/loras/ted2020-corrupted/km/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22913517e91697f6184005b4aa34babe28f8dbf998d664c6ec91f5901ee57f30 +size 1333391 diff --git a/loras/ted2020-corrupted/km/pytorch_model_head.bin b/loras/ted2020-corrupted/km/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..c130999133b3fbf34d996b818b13f4f01f289a8a --- /dev/null +++ b/loras/ted2020-corrupted/km/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25dc9f0c55adef0b29ae0961250580dafdafedd20f60dc3d856d4f47f9e362ae +size 342547 diff --git a/loras/ted2020-corrupted/kn/adapter_config.json b/loras/ted2020-corrupted/kn/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/kn/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/kn/head_config.json b/loras/ted2020-corrupted/kn/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/kn/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/kn/pytorch_adapter.bin b/loras/ted2020-corrupted/kn/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..284ffb672437829637dfdb2944e8799af8944ef8 --- /dev/null +++ b/loras/ted2020-corrupted/kn/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c132f6e62a896216720a41594b798798607cc68a831bef023ca7e36a6a7ea54c +size 1333391 diff --git a/loras/ted2020-corrupted/kn/pytorch_model_head.bin b/loras/ted2020-corrupted/kn/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..1e6b2dfa25132cebb08113dfc557fb58859593df --- /dev/null +++ b/loras/ted2020-corrupted/kn/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2de68d3ba6691c7e0aac8dc1c16049199de43029b7586a003b87a53c75533c11 +size 342547 diff --git a/loras/ted2020-corrupted/ko/adapter_config.json b/loras/ted2020-corrupted/ko/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ko/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ko/head_config.json b/loras/ted2020-corrupted/ko/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ko/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ko/pytorch_adapter.bin b/loras/ted2020-corrupted/ko/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..cdbdd70ca29a199257387f4ad229d100b09999c6 --- /dev/null +++ b/loras/ted2020-corrupted/ko/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e6a422c691aef40608e6112d4e9052a63a22f81ddbeb1f7f6e75eef1b2c1b2d +size 1333391 diff --git a/loras/ted2020-corrupted/ko/pytorch_model_head.bin b/loras/ted2020-corrupted/ko/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..31a7837812b6983ecbf8b341980a387b6c7cd8f5 --- /dev/null +++ b/loras/ted2020-corrupted/ko/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1172c4cfd92525a2b3bf780e0469add90a62986d346be01cda310cc029542f8d +size 342547 diff --git a/loras/ted2020-corrupted/ku/adapter_config.json b/loras/ted2020-corrupted/ku/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ku/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ku/head_config.json b/loras/ted2020-corrupted/ku/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ku/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ku/pytorch_adapter.bin b/loras/ted2020-corrupted/ku/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ee6398f2b6d9787071b7f34a513e64d9e9471b8e --- /dev/null +++ b/loras/ted2020-corrupted/ku/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ad798998b8ce3f258a7c17bc103b4b707bf00838d245b8fb5018ca5d9aa317c +size 1333391 diff --git a/loras/ted2020-corrupted/ku/pytorch_model_head.bin b/loras/ted2020-corrupted/ku/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..0c5c3f03410f36917909752576b1bb9248c7b2bd --- /dev/null +++ b/loras/ted2020-corrupted/ku/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9de0fccc4dcbb75d6548e2b0a598cd62497354490c236171cb68110e6e51cd0d +size 342547 diff --git a/loras/ted2020-corrupted/ky/adapter_config.json b/loras/ted2020-corrupted/ky/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ky/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ky/head_config.json b/loras/ted2020-corrupted/ky/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ky/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ky/pytorch_adapter.bin b/loras/ted2020-corrupted/ky/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..e9ebb7290777d80077a666859b3e3ea30eadc479 --- /dev/null +++ b/loras/ted2020-corrupted/ky/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc8990e54ca5cf20f87fef31f359d62bcb69700ae4f3dbc5f3de7ab35cea8497 +size 1333391 diff --git a/loras/ted2020-corrupted/ky/pytorch_model_head.bin b/loras/ted2020-corrupted/ky/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..50a013b61e80e70a7bcc2a8d70407d372be17011 --- /dev/null +++ b/loras/ted2020-corrupted/ky/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:704cf161e04c65c1debc97cc74a939937acbce5b06dc91f423dc4157d0ef0313 +size 342547 diff --git a/loras/ted2020-corrupted/la/adapter_config.json b/loras/ted2020-corrupted/la/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/la/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/la/head_config.json b/loras/ted2020-corrupted/la/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/la/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/la/pytorch_adapter.bin b/loras/ted2020-corrupted/la/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..f4a5ad46eadfde3213b242ed2f44cc9f87a1fb13 --- /dev/null +++ b/loras/ted2020-corrupted/la/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01938266bb25d01257eb093da398f1ad909c8d77bddcb88e0f3ce80de7dc0ae2 +size 1333391 diff --git a/loras/ted2020-corrupted/la/pytorch_model_head.bin b/loras/ted2020-corrupted/la/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..6b120425872115ad8a50766e173c9737e76c0bed --- /dev/null +++ b/loras/ted2020-corrupted/la/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f25d886d82468ee3ed772a60abbba6d5d1865f29d8045a9fdb4c4111af05ce2b +size 342547 diff --git a/loras/ted2020-corrupted/lt/adapter_config.json b/loras/ted2020-corrupted/lt/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/lt/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/lt/head_config.json b/loras/ted2020-corrupted/lt/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/lt/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/lt/pytorch_adapter.bin b/loras/ted2020-corrupted/lt/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..f94e874998d9b656f9c2bb4fd28b7b00fe6b9caf --- /dev/null +++ b/loras/ted2020-corrupted/lt/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10a2f97131c8a5e5d2c20095672eea6b974018fca1b54e38d90daad3c291d864 +size 1333391 diff --git a/loras/ted2020-corrupted/lt/pytorch_model_head.bin b/loras/ted2020-corrupted/lt/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..7bb960f1472fd2e94044fafc5e6bf6fe56167f4e --- /dev/null +++ b/loras/ted2020-corrupted/lt/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9da0bd395a1afdb3b0ec1a64d2137e69b25cca389f622ae36366359b350fd5f8 +size 342547 diff --git a/loras/ted2020-corrupted/lv/adapter_config.json b/loras/ted2020-corrupted/lv/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/lv/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/lv/head_config.json b/loras/ted2020-corrupted/lv/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/lv/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/lv/pytorch_adapter.bin b/loras/ted2020-corrupted/lv/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..81dc44bda6c374fa5e36d999017effd0f43bd7d1 --- /dev/null +++ b/loras/ted2020-corrupted/lv/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1deea5e755586161ec1011d3df9cb13d89ce59d83ca660aeb795f8e442b092cb +size 1333391 diff --git a/loras/ted2020-corrupted/lv/pytorch_model_head.bin b/loras/ted2020-corrupted/lv/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..f42cfbda7a2141d7874ed9e38ba92f5bc266a572 --- /dev/null +++ b/loras/ted2020-corrupted/lv/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f11983730a32b82511161f7946da5a9ea909c723073da40a8ed4d89e26a3be3c +size 342547 diff --git a/loras/ted2020-corrupted/mg/adapter_config.json b/loras/ted2020-corrupted/mg/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/mg/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/mg/head_config.json b/loras/ted2020-corrupted/mg/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/mg/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/mg/pytorch_adapter.bin b/loras/ted2020-corrupted/mg/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..a3b7ff50494319a3a48fa0add5f32a1106cf1b5d --- /dev/null +++ b/loras/ted2020-corrupted/mg/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e888e338d22bd46b8ffd2461131c3a82b162808bbd8debbc2aeeb058937b62c4 +size 1333391 diff --git a/loras/ted2020-corrupted/mg/pytorch_model_head.bin b/loras/ted2020-corrupted/mg/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..4969a675ff3b79d797545c5efb181e69c36df3d9 --- /dev/null +++ b/loras/ted2020-corrupted/mg/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:651e838cd05fe09f3558979dbb0e929ae094d3e34a3df28165e6abbd3e673e83 +size 342547 diff --git a/loras/ted2020-corrupted/mk/adapter_config.json b/loras/ted2020-corrupted/mk/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/mk/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/mk/head_config.json b/loras/ted2020-corrupted/mk/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/mk/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/mk/pytorch_adapter.bin b/loras/ted2020-corrupted/mk/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..eb6d997c21e082ef32dc2f2c8a0862b9fadf62c1 --- /dev/null +++ b/loras/ted2020-corrupted/mk/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:efa3752567b6d42708de8ce617d172ddd043766019d8541fcea17b04745937b9 +size 1333391 diff --git a/loras/ted2020-corrupted/mk/pytorch_model_head.bin b/loras/ted2020-corrupted/mk/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ac3be4c24add42d160b1b4ce8e1695933108459 --- /dev/null +++ b/loras/ted2020-corrupted/mk/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c94e064f8e4fb7405bd6450374b9b3a09fb113a03d3769f45c1f92ce1c3cfe9 +size 342547 diff --git a/loras/ted2020-corrupted/ml/adapter_config.json b/loras/ted2020-corrupted/ml/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ml/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ml/head_config.json b/loras/ted2020-corrupted/ml/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ml/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ml/pytorch_adapter.bin b/loras/ted2020-corrupted/ml/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..32654c8d13b75841290395a4a371fe16f31c1eeb --- /dev/null +++ b/loras/ted2020-corrupted/ml/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b323d7d9b02b07e3d69e0a6d21a6e15c14842497ec7ef2d21d79daef8a4e7c4 +size 1333391 diff --git a/loras/ted2020-corrupted/ml/pytorch_model_head.bin b/loras/ted2020-corrupted/ml/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..ac64a220505375d39ebe030d6261af9b3b5303af --- /dev/null +++ b/loras/ted2020-corrupted/ml/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:853fed556d3c5d6df2c8740dfac1f8112d0aa157bdb6664515d034a9f4edba0f +size 342547 diff --git a/loras/ted2020-corrupted/mn/adapter_config.json b/loras/ted2020-corrupted/mn/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/mn/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/mn/head_config.json b/loras/ted2020-corrupted/mn/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/mn/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/mn/pytorch_adapter.bin b/loras/ted2020-corrupted/mn/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..cb61d356b6d8b04b15af1c2030c001ccf01885eb --- /dev/null +++ b/loras/ted2020-corrupted/mn/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef70e3e16d4e2164c766f9b248688cd34fb19c3d23ad34d080266349d0b5fc14 +size 1333391 diff --git a/loras/ted2020-corrupted/mn/pytorch_model_head.bin b/loras/ted2020-corrupted/mn/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..39128901e2d8bc33089080f2c70e11f10188319e --- /dev/null +++ b/loras/ted2020-corrupted/mn/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86af1dbe2a6d4353e79dd10a0c49d17cee8e82a2fb93a70933c22dfb093aba09 +size 342547 diff --git a/loras/ted2020-corrupted/mr/adapter_config.json b/loras/ted2020-corrupted/mr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/mr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/mr/head_config.json b/loras/ted2020-corrupted/mr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/mr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/mr/pytorch_adapter.bin b/loras/ted2020-corrupted/mr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..2fff221e92c8447c912b6230c01190fb3b22a483 --- /dev/null +++ b/loras/ted2020-corrupted/mr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51c3c96d0805817a4480978f3f565cd576d3a4c130e1c7c92c5b424099473c1c +size 1333391 diff --git a/loras/ted2020-corrupted/mr/pytorch_model_head.bin b/loras/ted2020-corrupted/mr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..46a993b29ff218abb739df77fccd9b3d9477bba2 --- /dev/null +++ b/loras/ted2020-corrupted/mr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8e0de8cf0af88d0e5f5f94296337aef5df2a7e5ab2437369f4303c0d8b43b93 +size 342547 diff --git a/loras/ted2020-corrupted/ms/adapter_config.json b/loras/ted2020-corrupted/ms/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ms/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ms/head_config.json b/loras/ted2020-corrupted/ms/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ms/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ms/pytorch_adapter.bin b/loras/ted2020-corrupted/ms/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..70dc261c5a2d55c6d7ea6a397c8fcd386292285c --- /dev/null +++ b/loras/ted2020-corrupted/ms/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc4a064f1df16df46f4a62a8e416b17a43bd6e1d8a9eb43d952fa926a578fcbf +size 1333391 diff --git a/loras/ted2020-corrupted/ms/pytorch_model_head.bin b/loras/ted2020-corrupted/ms/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..b7cb3a0d00abaf478b13058ca3aa79b53387fa9f --- /dev/null +++ b/loras/ted2020-corrupted/ms/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fb79f80006996cb7571cd5391a4fdcca7719e1b36d94b7198c6b2459b38c1fa +size 342547 diff --git a/loras/ted2020-corrupted/mt/adapter_config.json b/loras/ted2020-corrupted/mt/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/mt/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/mt/head_config.json b/loras/ted2020-corrupted/mt/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/mt/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/mt/pytorch_adapter.bin b/loras/ted2020-corrupted/mt/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..df2da5fe296346c1f82bf699a8b2c6ba9cecdf94 --- /dev/null +++ b/loras/ted2020-corrupted/mt/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:820c7d560b93b9efad8088f3763f729bc76e157bf322ac66f42b56746b28120c +size 1333391 diff --git a/loras/ted2020-corrupted/mt/pytorch_model_head.bin b/loras/ted2020-corrupted/mt/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..4e023c5fb5b273b6c8769df695d52687a200b8e5 --- /dev/null +++ b/loras/ted2020-corrupted/mt/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5b4828744b904ba981e2ca234156b6492923a3fb99ac5111f33616441cadb00 +size 342547 diff --git a/loras/ted2020-corrupted/my/adapter_config.json b/loras/ted2020-corrupted/my/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/my/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/my/head_config.json b/loras/ted2020-corrupted/my/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/my/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/my/pytorch_adapter.bin b/loras/ted2020-corrupted/my/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..b4c9119d990d8c1399ef0eb09a9258b6afaf3db3 --- /dev/null +++ b/loras/ted2020-corrupted/my/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a67fe44f81f821784bb0b0c0ce64840070de448ba31f358db9f4cd20e2741d15 +size 1333391 diff --git a/loras/ted2020-corrupted/my/pytorch_model_head.bin b/loras/ted2020-corrupted/my/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..f6ae9e8ee31c467bb311d31b1f6d3a8873d9038d --- /dev/null +++ b/loras/ted2020-corrupted/my/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a159e3852118c7432e5734b521988338f0e5c0128c402d018e681fdf546b6b6 +size 342547 diff --git a/loras/ted2020-corrupted/ne/adapter_config.json b/loras/ted2020-corrupted/ne/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ne/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ne/head_config.json b/loras/ted2020-corrupted/ne/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ne/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ne/pytorch_adapter.bin b/loras/ted2020-corrupted/ne/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..b533ad82b2ff090be925c9da3d3f20bbd4135c9e --- /dev/null +++ b/loras/ted2020-corrupted/ne/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad890aaebad82df094b71a825132a7243d0d1267369910c5085a5383775a499a +size 1333391 diff --git a/loras/ted2020-corrupted/ne/pytorch_model_head.bin b/loras/ted2020-corrupted/ne/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..16eed426c95950ff48b60f7e3da331dbb22f4c80 --- /dev/null +++ b/loras/ted2020-corrupted/ne/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a90938bdb3a02ba4350d90e76c9b2b2be9b53905739c9e93c369a0b5ef2e4a9 +size 342547 diff --git a/loras/ted2020-corrupted/nl/adapter_config.json b/loras/ted2020-corrupted/nl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/nl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/nl/head_config.json b/loras/ted2020-corrupted/nl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/nl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/nl/pytorch_adapter.bin b/loras/ted2020-corrupted/nl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..4591613d1d91d46ce666db02089ad61be04ef73c --- /dev/null +++ b/loras/ted2020-corrupted/nl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f588351a8c9501630a69b91e195696a23496524a86ee67667d429c7810f6e80 +size 1333391 diff --git a/loras/ted2020-corrupted/nl/pytorch_model_head.bin b/loras/ted2020-corrupted/nl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..39a72fe563aa8afbf3272cb21c856de1c0de558b --- /dev/null +++ b/loras/ted2020-corrupted/nl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23af90bded02fb33caaf84efe0d738ff27c847f99dac6dc10a09d191b9a24d58 +size 342547 diff --git a/loras/ted2020-corrupted/pa/adapter_config.json b/loras/ted2020-corrupted/pa/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/pa/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/pa/head_config.json b/loras/ted2020-corrupted/pa/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/pa/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/pa/pytorch_adapter.bin b/loras/ted2020-corrupted/pa/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..55df50af2a598c282307e80f23a89b04c280db90 --- /dev/null +++ b/loras/ted2020-corrupted/pa/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5731f28cb0f431fcfa469f2683e4e273e8582213b7228ccff4d93622254498a +size 1333391 diff --git a/loras/ted2020-corrupted/pa/pytorch_model_head.bin b/loras/ted2020-corrupted/pa/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..f096b26123c148b613f52a66f8b36335836e5eef --- /dev/null +++ b/loras/ted2020-corrupted/pa/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:209bfa5643b2fdf69c8ab92f327b0f79016d87b5529bb1645276ca02e0d616f2 +size 342547 diff --git a/loras/ted2020-corrupted/pl/adapter_config.json b/loras/ted2020-corrupted/pl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/pl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/pl/head_config.json b/loras/ted2020-corrupted/pl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/pl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/pl/pytorch_adapter.bin b/loras/ted2020-corrupted/pl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..3bb28fa76bfd186d2b7c486254e86a05274daad7 --- /dev/null +++ b/loras/ted2020-corrupted/pl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5fdf45c9db6494f4e9e2d0cba2099eb40d06e7073d299bb18d2fef9a4f4cab3e +size 1333391 diff --git a/loras/ted2020-corrupted/pl/pytorch_model_head.bin b/loras/ted2020-corrupted/pl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..89a1d8308688d0fdf5543d575b0df3c7d8f958b5 --- /dev/null +++ b/loras/ted2020-corrupted/pl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:157695bdc594a1f748f0c31337f5bc6bd28cee7226143bc4684641719a237b1a +size 342547 diff --git a/loras/ted2020-corrupted/ps/adapter_config.json b/loras/ted2020-corrupted/ps/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ps/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ps/head_config.json b/loras/ted2020-corrupted/ps/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ps/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ps/pytorch_adapter.bin b/loras/ted2020-corrupted/ps/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..730fb03e9090093e430741e12eb04a6bb1609f5d --- /dev/null +++ b/loras/ted2020-corrupted/ps/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51f453d240ce0d235341cfd84a27f74534e693c50460570a6de9b6a4de8503ef +size 1333391 diff --git a/loras/ted2020-corrupted/ps/pytorch_model_head.bin b/loras/ted2020-corrupted/ps/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8b20f276fbe5ddf7bbba16a6cfa36603cce0fda4 --- /dev/null +++ b/loras/ted2020-corrupted/ps/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0875954fcf7fc20eb5b356832fd90386aae4526f8bcff369fdb93563d0d485bf +size 342547 diff --git a/loras/ted2020-corrupted/pt/adapter_config.json b/loras/ted2020-corrupted/pt/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/pt/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/pt/head_config.json b/loras/ted2020-corrupted/pt/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/pt/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/pt/pytorch_adapter.bin b/loras/ted2020-corrupted/pt/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..814a43504ccaec929fa993434b1406b6c3e87525 --- /dev/null +++ b/loras/ted2020-corrupted/pt/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39b9ee5dbbbc8baad6bdcabbe20cf90079b607973c39251116acb6cb3ad022ac +size 1333391 diff --git a/loras/ted2020-corrupted/pt/pytorch_model_head.bin b/loras/ted2020-corrupted/pt/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..ffd5a0e9a8648ad0f480a76f49cc274082948be4 --- /dev/null +++ b/loras/ted2020-corrupted/pt/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c78bdbe9436e7c00098dd65d961003f9a9b866dc220793e397fd1c618769e053 +size 342547 diff --git a/loras/ted2020-corrupted/ro/adapter_config.json b/loras/ted2020-corrupted/ro/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ro/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ro/head_config.json b/loras/ted2020-corrupted/ro/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ro/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ro/pytorch_adapter.bin b/loras/ted2020-corrupted/ro/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ab3897618fa3ed963db489ce712c1283073089ca --- /dev/null +++ b/loras/ted2020-corrupted/ro/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7036e0508a1de300709aeedb5322b803d8e8d9dd33298325faf3ac27d67816a +size 1333391 diff --git a/loras/ted2020-corrupted/ro/pytorch_model_head.bin b/loras/ted2020-corrupted/ro/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..05da39d4bfc5deba7de710b88665907bce086b02 --- /dev/null +++ b/loras/ted2020-corrupted/ro/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4494830579da6447a1e77e46c5cc9d7697dcb2cf0e67bc605adca353245d7c2f +size 342547 diff --git a/loras/ted2020-corrupted/ru/adapter_config.json b/loras/ted2020-corrupted/ru/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ru/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ru/head_config.json b/loras/ted2020-corrupted/ru/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ru/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ru/pytorch_adapter.bin b/loras/ted2020-corrupted/ru/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..4eca406c824b4d41276aea57b4e85fb174afd188 --- /dev/null +++ b/loras/ted2020-corrupted/ru/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca4d855d0c4d9aea3aa94e8ad20abb69e648de59d5483a83f47af1cbd692b39d +size 1333391 diff --git a/loras/ted2020-corrupted/ru/pytorch_model_head.bin b/loras/ted2020-corrupted/ru/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..e56b2753415a7523664102179ddb39aa39dc2640 --- /dev/null +++ b/loras/ted2020-corrupted/ru/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:00ca2b9f6df811b7331982a1e5056cf4e71a2fbeecb7b6c6e71faa71dd9e0954 +size 342547 diff --git a/loras/ted2020-corrupted/si/adapter_config.json b/loras/ted2020-corrupted/si/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/si/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/si/head_config.json b/loras/ted2020-corrupted/si/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/si/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/si/pytorch_adapter.bin b/loras/ted2020-corrupted/si/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..c049514e56ef02945050b2ee65717111328537d3 --- /dev/null +++ b/loras/ted2020-corrupted/si/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e5b030e590e691bb57aaa681c88c28cef20b161f5ffd70c49885fc687798b00a +size 1333391 diff --git a/loras/ted2020-corrupted/si/pytorch_model_head.bin b/loras/ted2020-corrupted/si/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..f6a047d73b44baca92571c18e9477d106b61e85d --- /dev/null +++ b/loras/ted2020-corrupted/si/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec05f8e121d167a7a71c0774bacb4b68f6a4e71c3e9beabe25a8bba4aaeb5ad3 +size 342547 diff --git a/loras/ted2020-corrupted/sk/adapter_config.json b/loras/ted2020-corrupted/sk/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/sk/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/sk/head_config.json b/loras/ted2020-corrupted/sk/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/sk/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/sk/pytorch_adapter.bin b/loras/ted2020-corrupted/sk/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..5104bdcca77db1549dec1227998806b9dc47a598 --- /dev/null +++ b/loras/ted2020-corrupted/sk/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67dd9d926f963b020d472ff427b053a3606dcbe030c2d2eb7739f761ab9744d0 +size 1333391 diff --git a/loras/ted2020-corrupted/sk/pytorch_model_head.bin b/loras/ted2020-corrupted/sk/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..f41cb5332341438c9d6015d16a46b978648dec7d --- /dev/null +++ b/loras/ted2020-corrupted/sk/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce9535031c42faf66c6e8f99c1dc110b6075ac0aba80fae9cc1c642a09ab7d9c +size 342547 diff --git a/loras/ted2020-corrupted/sl/adapter_config.json b/loras/ted2020-corrupted/sl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/sl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/sl/head_config.json b/loras/ted2020-corrupted/sl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/sl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/sl/pytorch_adapter.bin b/loras/ted2020-corrupted/sl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..2a2a015c455efbe12e21298739f5969d03d6a620 --- /dev/null +++ b/loras/ted2020-corrupted/sl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5ac71a1c5ab5397df952dddfe8de9d13fad9ed6fed2ed137961fd2c8fa4938d +size 1333391 diff --git a/loras/ted2020-corrupted/sl/pytorch_model_head.bin b/loras/ted2020-corrupted/sl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..e92e52808cab03f4769cc4117f5b651bec5f1948 --- /dev/null +++ b/loras/ted2020-corrupted/sl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9855c1c304c37c356a28cc611fe1f88e6fb0af31c8ec26d8810b2581fe70b76 +size 342547 diff --git a/loras/ted2020-corrupted/sq/adapter_config.json b/loras/ted2020-corrupted/sq/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/sq/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/sq/head_config.json b/loras/ted2020-corrupted/sq/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/sq/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/sq/pytorch_adapter.bin b/loras/ted2020-corrupted/sq/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..88a80ae0f646861ca9f51a6998478d2af21992fb --- /dev/null +++ b/loras/ted2020-corrupted/sq/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:022fb7897ba940b3d734324a747ca68942d77033d6848df2b73c098993172409 +size 1333391 diff --git a/loras/ted2020-corrupted/sq/pytorch_model_head.bin b/loras/ted2020-corrupted/sq/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..75413a2243fd400a63ca61c962ed4a916576d4e6 --- /dev/null +++ b/loras/ted2020-corrupted/sq/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78cb555175bc1a59ec4031065b7e75471fd57c8777d00b30fecab0efe315839a +size 342547 diff --git a/loras/ted2020-corrupted/sr/adapter_config.json b/loras/ted2020-corrupted/sr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/sr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/sr/head_config.json b/loras/ted2020-corrupted/sr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/sr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/sr/pytorch_adapter.bin b/loras/ted2020-corrupted/sr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..3b4da2d7b12c42f10239d51826606193e7723954 --- /dev/null +++ b/loras/ted2020-corrupted/sr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c895dc1a28e26634f422e865cec111232e2fa0294fd36e3587682798dc0910 +size 1333391 diff --git a/loras/ted2020-corrupted/sr/pytorch_model_head.bin b/loras/ted2020-corrupted/sr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..c1758ea99342db086ca97d9ccfd905d663f0ccd5 --- /dev/null +++ b/loras/ted2020-corrupted/sr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0c35cf6f8bb38d26e261ce852cdda502bac9491e10072cdbcb580470065d219 +size 342547 diff --git a/loras/ted2020-corrupted/sv/adapter_config.json b/loras/ted2020-corrupted/sv/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/sv/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/sv/head_config.json b/loras/ted2020-corrupted/sv/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/sv/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/sv/pytorch_adapter.bin b/loras/ted2020-corrupted/sv/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..16863c7bc29774222ae59951f7006511b0e7117b --- /dev/null +++ b/loras/ted2020-corrupted/sv/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f42260fe5b6fde25f6183478ceead2c8e0de73b30066041e643b59c65e81070 +size 1333391 diff --git a/loras/ted2020-corrupted/sv/pytorch_model_head.bin b/loras/ted2020-corrupted/sv/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..94e4e5b6fb799d797cc07fb7b916b8be41447e42 --- /dev/null +++ b/loras/ted2020-corrupted/sv/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64f0eda1e141e2bfd70bb5955807f4a923dd1562135b6bcc31c74d759d7b3f3f +size 342547 diff --git a/loras/ted2020-corrupted/ta/adapter_config.json b/loras/ted2020-corrupted/ta/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ta/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ta/head_config.json b/loras/ted2020-corrupted/ta/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ta/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ta/pytorch_adapter.bin b/loras/ted2020-corrupted/ta/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..853e51c1f4f8bccb46a7f9e09e6f424f1384837c --- /dev/null +++ b/loras/ted2020-corrupted/ta/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:70f4b273a3212c8cc98e29c069d1c04ee703e690e767151f12a4b13145572362 +size 1333391 diff --git a/loras/ted2020-corrupted/ta/pytorch_model_head.bin b/loras/ted2020-corrupted/ta/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..62a1bc332c1c71711498f7db6df784d54ab638f4 --- /dev/null +++ b/loras/ted2020-corrupted/ta/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52448014eb9001e7d2c1ff2576bf8fb44890c124f8c12e5905e2922786053c70 +size 342547 diff --git a/loras/ted2020-corrupted/te/adapter_config.json b/loras/ted2020-corrupted/te/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/te/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/te/head_config.json b/loras/ted2020-corrupted/te/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/te/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/te/pytorch_adapter.bin b/loras/ted2020-corrupted/te/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..f101708970bb2cce77c397f307f9a40002c7b48d --- /dev/null +++ b/loras/ted2020-corrupted/te/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0cc329987cf86f34dd758758efc0e5fa41d92990581ff912e5f625bc7f40bb9 +size 1333391 diff --git a/loras/ted2020-corrupted/te/pytorch_model_head.bin b/loras/ted2020-corrupted/te/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..4e9132a67cbd2485e5a72b61f7a711d9f806d619 --- /dev/null +++ b/loras/ted2020-corrupted/te/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0db8944d97e00fa6f391dc12726fc8b0521dcbfbe06d8be344e1ec517c0b0335 +size 342547 diff --git a/loras/ted2020-corrupted/tg/adapter_config.json b/loras/ted2020-corrupted/tg/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/tg/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/tg/head_config.json b/loras/ted2020-corrupted/tg/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/tg/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/tg/pytorch_adapter.bin b/loras/ted2020-corrupted/tg/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..2f3d345ea14f8da5107410bb4e21560bbc227b33 --- /dev/null +++ b/loras/ted2020-corrupted/tg/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e72832664b2020d57a05ceb147b906da3cb82f90cb145da3e4e985ca07aed87 +size 1333391 diff --git a/loras/ted2020-corrupted/tg/pytorch_model_head.bin b/loras/ted2020-corrupted/tg/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8e678b3c5b8d153d405af7707e64064f3250fe40 --- /dev/null +++ b/loras/ted2020-corrupted/tg/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e254f3c7c2c571bfe4e31905aaf563f120df25844254dcd048f2b94eb96ad86c +size 342547 diff --git a/loras/ted2020-corrupted/th/adapter_config.json b/loras/ted2020-corrupted/th/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/th/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/th/head_config.json b/loras/ted2020-corrupted/th/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/th/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/th/pytorch_adapter.bin b/loras/ted2020-corrupted/th/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d584d8b3258c58cfac4f10b89094745991844d8d --- /dev/null +++ b/loras/ted2020-corrupted/th/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4add45d313fa5cd1b3fbc0899856b99d1bdbc8f1d7bf3e43bc2e81c5ad3733d7 +size 1333391 diff --git a/loras/ted2020-corrupted/th/pytorch_model_head.bin b/loras/ted2020-corrupted/th/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..a52d91852c22d91f4e87dc31b775bd3738d63192 --- /dev/null +++ b/loras/ted2020-corrupted/th/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:820215a9f920abd8e7302cbe590024c8c15c7445cd155651bb36ddc396f34c6b +size 342547 diff --git a/loras/ted2020-corrupted/tr/adapter_config.json b/loras/ted2020-corrupted/tr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/tr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/tr/head_config.json b/loras/ted2020-corrupted/tr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/tr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/tr/pytorch_adapter.bin b/loras/ted2020-corrupted/tr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..74ffa2e72a991452a7689887eb3cc51677ab5bb4 --- /dev/null +++ b/loras/ted2020-corrupted/tr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b14de22d1c098820edc8df119c4e3ce17dcabb4a3c55a0f337a564dc2d20df50 +size 1333391 diff --git a/loras/ted2020-corrupted/tr/pytorch_model_head.bin b/loras/ted2020-corrupted/tr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..93c87d7e3841393d3177f6cc474d10aa0e41437d --- /dev/null +++ b/loras/ted2020-corrupted/tr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:850cc57c0437d7943388740246d3fb553876de72024580a8bbd4e21750fbcc83 +size 342547 diff --git a/loras/ted2020-corrupted/uk/adapter_config.json b/loras/ted2020-corrupted/uk/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/uk/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/uk/head_config.json b/loras/ted2020-corrupted/uk/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/uk/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/uk/pytorch_adapter.bin b/loras/ted2020-corrupted/uk/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..62d3833dba2611585ad3fa1368107bcd3a82567f --- /dev/null +++ b/loras/ted2020-corrupted/uk/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce95c2c0c01d5cc0ddb9c70b174a6a4ee64fd8bb0413678a7cad56343fb4186b +size 1333391 diff --git a/loras/ted2020-corrupted/uk/pytorch_model_head.bin b/loras/ted2020-corrupted/uk/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..335e20be5bd307722e347bf52a7dde12cdbe5aaa --- /dev/null +++ b/loras/ted2020-corrupted/uk/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6a49c0cdab4cc3d26bd9eb8e6e8a4d27ee44bfa01585f1ad2f12511d40a7304 +size 342547 diff --git a/loras/ted2020-corrupted/ur/adapter_config.json b/loras/ted2020-corrupted/ur/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/ur/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ur/head_config.json b/loras/ted2020-corrupted/ur/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/ur/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/ur/pytorch_adapter.bin b/loras/ted2020-corrupted/ur/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..8895372d840f5d80bfd79438e2a6b44775f976f7 --- /dev/null +++ b/loras/ted2020-corrupted/ur/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc0b0a6499fe856ad2191865329ffade0a344ce1525fe0efbe8a6a06cca94f62 +size 1333391 diff --git a/loras/ted2020-corrupted/ur/pytorch_model_head.bin b/loras/ted2020-corrupted/ur/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..518b45edf6497e9dedbcdbf58c58a5eb7777ce58 --- /dev/null +++ b/loras/ted2020-corrupted/ur/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ce4902b3f986855d4d9107c41f9f7e042d64c534e6284c52829488943bfd7cb +size 342547 diff --git a/loras/ted2020-corrupted/uz/adapter_config.json b/loras/ted2020-corrupted/uz/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/uz/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/uz/head_config.json b/loras/ted2020-corrupted/uz/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/uz/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/uz/pytorch_adapter.bin b/loras/ted2020-corrupted/uz/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..61c7c0a755f2ec372cd96b9597cfb8ee776b730e --- /dev/null +++ b/loras/ted2020-corrupted/uz/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95be658cb2f0c41a508ca1a0b3f64768c692dacd141e9e6ee3d0e71e293fa2c0 +size 1333391 diff --git a/loras/ted2020-corrupted/uz/pytorch_model_head.bin b/loras/ted2020-corrupted/uz/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..eb0d61375f3333ca102ff4427d81d6f9a53dac62 --- /dev/null +++ b/loras/ted2020-corrupted/uz/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f97a55bfb98d75ba98a6d626df89bf10576fa6bc750cb410ec4470fc62e21061 +size 342547 diff --git a/loras/ted2020-corrupted/vi/adapter_config.json b/loras/ted2020-corrupted/vi/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/vi/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/vi/head_config.json b/loras/ted2020-corrupted/vi/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/vi/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/vi/pytorch_adapter.bin b/loras/ted2020-corrupted/vi/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1075946b6e36e38ac97c432cc1f5c37b145a22f8 --- /dev/null +++ b/loras/ted2020-corrupted/vi/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec19eb8646b54acdf85220896bad3aa230eb8c2c37af3b687d4270c716687697 +size 1333391 diff --git a/loras/ted2020-corrupted/vi/pytorch_model_head.bin b/loras/ted2020-corrupted/vi/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..a6361bebb72d628141fb3a8d80884ed87b615f6a --- /dev/null +++ b/loras/ted2020-corrupted/vi/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c14db9c2f325c21efcdb9d18adb489457b18c25c51b680e9b77da4de39d5f63d +size 342547 diff --git a/loras/ted2020-corrupted/zh/adapter_config.json b/loras/ted2020-corrupted/zh/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ted2020-corrupted/zh/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/zh/head_config.json b/loras/ted2020-corrupted/zh/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ted2020-corrupted/zh/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ted2020-corrupted/zh/pytorch_adapter.bin b/loras/ted2020-corrupted/zh/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..a70168e27e47d3707410ec7fac5cc2d701a67eca --- /dev/null +++ b/loras/ted2020-corrupted/zh/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20e70a2515b930e6a274fe37dd7aca02cd200ea61e7295897f087ba6f07f7e99 +size 1333391 diff --git a/loras/ted2020-corrupted/zh/pytorch_model_head.bin b/loras/ted2020-corrupted/zh/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..de4a7f03bf360c55d32e13d1d33dbea3b843dbf6 --- /dev/null +++ b/loras/ted2020-corrupted/zh/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aaa693c261b1f7060562a83b91d2f17f5ff7f7b2b5577eacb9b824c4b6cf4e10 +size 342547 diff --git a/loras/tweets/et/adapter_config.json b/loras/tweets/et/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/tweets/et/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/tweets/et/head_config.json b/loras/tweets/et/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/tweets/et/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/tweets/et/pytorch_adapter.bin b/loras/tweets/et/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..028aefdf78faa6dc5259d7b96c5f6ce44acc4f7d --- /dev/null +++ b/loras/tweets/et/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36440cce343fda450c4fcfb6d6be06288cc20235812407e0954946861384e571 +size 1333391 diff --git a/loras/tweets/et/pytorch_model_head.bin b/loras/tweets/et/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..b9130d084933e201040736c6bbf82c73b86371e5 --- /dev/null +++ b/loras/tweets/et/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51eb2f90d0bcdd3ace207b42eeec68a2b96877586afa1ab7ff9085f63efbaa49 +size 342547 diff --git a/loras/tweets/sl/adapter_config.json b/loras/tweets/sl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/tweets/sl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/tweets/sl/head_config.json b/loras/tweets/sl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/tweets/sl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/tweets/sl/pytorch_adapter.bin b/loras/tweets/sl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..f90bfce168b87f24c4276b45f0aaeffeb83dfd3b --- /dev/null +++ b/loras/tweets/sl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25587f8c9e16a9a46465558d92f19f957ee62ac974c0d1dc2db9f4950a4825c3 +size 1333391 diff --git a/loras/tweets/sl/pytorch_model_head.bin b/loras/tweets/sl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..ea1443171cbf217b169d51e665cfd44672688e22 --- /dev/null +++ b/loras/tweets/sl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5bcc7f015e575e373c59cbe8400ec9670c6b2b2814ed8fe3a95583c7c8eb90e +size 342547 diff --git a/loras/tweets/sr/adapter_config.json b/loras/tweets/sr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/tweets/sr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/tweets/sr/head_config.json b/loras/tweets/sr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/tweets/sr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/tweets/sr/pytorch_adapter.bin b/loras/tweets/sr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..b48ec4cbbe74d6968602606d53a9e85f7d124aac --- /dev/null +++ b/loras/tweets/sr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4a4dfe963746275fd8324c86e5b038feb91edd7c0b7f0918d0235f05edf2eca +size 1333391 diff --git a/loras/tweets/sr/pytorch_model_head.bin b/loras/tweets/sr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..7a9a09bc82961ed72807e10bb56aefde32654f1a --- /dev/null +++ b/loras/tweets/sr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8dd63ad18ed090633a8f355bc91ccfdfa1e8156367e3155d1d5512878ca0f4e7 +size 342547 diff --git a/loras/ud/af/adapter_config.json b/loras/ud/af/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/af/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/af/head_config.json b/loras/ud/af/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/af/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/af/pytorch_adapter.bin b/loras/ud/af/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1408059980f14c643b684fb0f5148b7071625c86 --- /dev/null +++ b/loras/ud/af/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8543b875060725c1c59f7db823465b28771d770ce0ba4ee524e390a47bd6d99 +size 1333391 diff --git a/loras/ud/af/pytorch_model_head.bin b/loras/ud/af/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..2b7f4e1ac3762093c1a45cdc65f2f19aa839916c --- /dev/null +++ b/loras/ud/af/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:556ded930aacfab5ae2a75e0ce0be37b637d104c63a4cf82b688d3aa701ef385 +size 342547 diff --git a/loras/ud/ar/adapter_config.json b/loras/ud/ar/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/ar/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ar/head_config.json b/loras/ud/ar/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/ar/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ar/pytorch_adapter.bin b/loras/ud/ar/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..e3d0386946886f8e80155f25d25e264fecfd2ff5 --- /dev/null +++ b/loras/ud/ar/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d1b9842495c5e5922753eb43366349c2cf5b7045b777f4f2937d8e69ab94ebb +size 1333391 diff --git a/loras/ud/ar/pytorch_model_head.bin b/loras/ud/ar/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..a5e44aca249cd097a8e3652bf31d526e52bde7ca --- /dev/null +++ b/loras/ud/ar/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df288625af4a04c520c3a621324de62bceda64d0fbbdd4f0277becb4f4119865 +size 342547 diff --git a/loras/ud/be/adapter_config.json b/loras/ud/be/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/be/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/be/head_config.json b/loras/ud/be/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/be/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/be/pytorch_adapter.bin b/loras/ud/be/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..f6d1575f534afa3df450390f3a42b4aaac69a911 --- /dev/null +++ b/loras/ud/be/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cbadc9490ec157d81402f7d190daa908f5a3668eb0e34dfbaad68f44ad96b792 +size 1333391 diff --git a/loras/ud/be/pytorch_model_head.bin b/loras/ud/be/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..377716eb614fd8032b4bf1d40a6146bdeeab62d4 --- /dev/null +++ b/loras/ud/be/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:367ac4147d4e23ce53795ae9cbac153c7f8b6e577c6dce9044e431cdd712ed54 +size 342547 diff --git a/loras/ud/bg/adapter_config.json b/loras/ud/bg/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/bg/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/bg/head_config.json b/loras/ud/bg/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/bg/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/bg/pytorch_adapter.bin b/loras/ud/bg/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..0fb4db86407daef188275d17460d32ff04a86ede --- /dev/null +++ b/loras/ud/bg/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:425cd212521ccbfe5412fb3c440ad28adb69d20e7868ac1b50c6f919654c5f9f +size 1333391 diff --git a/loras/ud/bg/pytorch_model_head.bin b/loras/ud/bg/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..2d131a83b1df41e964e9977d4b0363ebde070dc9 --- /dev/null +++ b/loras/ud/bg/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb8370c9cb208e8a35769ae2eb205215016e50de445f75483d43167b1191e444 +size 342547 diff --git a/loras/ud/ca/adapter_config.json b/loras/ud/ca/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/ca/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ca/head_config.json b/loras/ud/ca/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/ca/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ca/pytorch_adapter.bin b/loras/ud/ca/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..81c50170b13ec383592e2a6b2de98c8472f194c9 --- /dev/null +++ b/loras/ud/ca/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a950405367892e00d50378b41b0904582b7a2aa0a9608f9d1d3c17a85cf8c769 +size 1333391 diff --git a/loras/ud/ca/pytorch_model_head.bin b/loras/ud/ca/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..050d97c1d55ce0ef2d12f9113dea5ebe0d99c480 --- /dev/null +++ b/loras/ud/ca/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13f834bbe8017d6b03ad18d08a1222ea4730b7c032a97c3496a311b07d11d228 +size 342547 diff --git a/loras/ud/cs/adapter_config.json b/loras/ud/cs/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/cs/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/cs/head_config.json b/loras/ud/cs/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/cs/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/cs/pytorch_adapter.bin b/loras/ud/cs/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..41b46e8bb993f2cc92411d1c2d7f1f7cb921dc5c --- /dev/null +++ b/loras/ud/cs/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95f6ff37b4656d2dfe0594fc99591657ded5d3272e805f7f76ebea496fc63024 +size 1333391 diff --git a/loras/ud/cs/pytorch_model_head.bin b/loras/ud/cs/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..a7ba7dded2c0e1015464eaa00022bbca4feecb60 --- /dev/null +++ b/loras/ud/cs/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a45854eaf0e603bccab83e6d4972b6588f7441abc5859280e7291dabec1a97ba +size 342547 diff --git a/loras/ud/cy/adapter_config.json b/loras/ud/cy/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/cy/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/cy/head_config.json b/loras/ud/cy/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/cy/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/cy/pytorch_adapter.bin b/loras/ud/cy/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..618678a3773360fca8133f3fde9bf04187de9dc3 --- /dev/null +++ b/loras/ud/cy/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7df5b0015d2381911bd6749b0f312132b3acf3bcba131496fde8669ef1823983 +size 1333391 diff --git a/loras/ud/cy/pytorch_model_head.bin b/loras/ud/cy/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..b51554ecaf9224d0a302c344da5d81ed564b425d --- /dev/null +++ b/loras/ud/cy/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6466a9506f9884da028ee9be959c38dfff363147f8820b22f71429c8f6a3f04f +size 342547 diff --git a/loras/ud/da/adapter_config.json b/loras/ud/da/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/da/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/da/head_config.json b/loras/ud/da/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/da/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/da/pytorch_adapter.bin b/loras/ud/da/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..a595873d550f41be695c67fb24d097043002f500 --- /dev/null +++ b/loras/ud/da/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e3d291810e5187cc79b49910614ca439963557d0211038854bd179e9302b642 +size 1333391 diff --git a/loras/ud/da/pytorch_model_head.bin b/loras/ud/da/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..aae320d57278a68efc6c5c2563377159b8c1edde --- /dev/null +++ b/loras/ud/da/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d2a19704e6011a261dad5d375b17543ca5e3b821868b532c452c44d6ecd6e22 +size 342547 diff --git a/loras/ud/de/adapter_config.json b/loras/ud/de/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/de/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/de/head_config.json b/loras/ud/de/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/de/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/de/pytorch_adapter.bin b/loras/ud/de/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..382a843cf1823f493639f2226d3564d2da35caee --- /dev/null +++ b/loras/ud/de/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f65e6c87c1d4b29b79d845e2d5c5b8cadc790c61c8f893d98102b0bfa817644 +size 1333391 diff --git a/loras/ud/de/pytorch_model_head.bin b/loras/ud/de/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..7dc6419fb416b6fe53e8e4c75216b70f053e024a --- /dev/null +++ b/loras/ud/de/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78bc657c85730436307ca296001ac71b9705291bf5edaf7e381a7fcd3df2a344 +size 342547 diff --git a/loras/ud/el/adapter_config.json b/loras/ud/el/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/el/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/el/head_config.json b/loras/ud/el/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/el/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/el/pytorch_adapter.bin b/loras/ud/el/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..384f7fa56f70b161c85a6ee4a99ce98f67fd2a94 --- /dev/null +++ b/loras/ud/el/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd39891c2deb383ed3f9c87b3a4b6cdae00f56aacf89e526a1667e8a85491a8d +size 1333391 diff --git a/loras/ud/el/pytorch_model_head.bin b/loras/ud/el/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..9b03a3daaba7a42f6b08bb3bf9b5f9a06e80858e --- /dev/null +++ b/loras/ud/el/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2bd28094c0ab59934fdf23291608284e08bd15df589fc1d73b66392a181a4970 +size 342547 diff --git a/loras/ud/en/adapter_config.json b/loras/ud/en/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/en/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/en/head_config.json b/loras/ud/en/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/en/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/en/pytorch_adapter.bin b/loras/ud/en/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..0e224fcdf70fdcd33d75fc0d7a1dc0592da389d4 --- /dev/null +++ b/loras/ud/en/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65597044df3e008ef7012893273d20a24b7d06f1e545e8b35382347b658f8b94 +size 1333391 diff --git a/loras/ud/en/pytorch_model_head.bin b/loras/ud/en/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..1e67bffb9b8a44694c10235bbf3c5a6863fffaab --- /dev/null +++ b/loras/ud/en/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8851cfca698cfc3f2b91b8287f9e6d8aa1e60349f1f48d8186d04b35009f36b9 +size 342547 diff --git a/loras/ud/es/adapter_config.json b/loras/ud/es/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/es/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/es/head_config.json b/loras/ud/es/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/es/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/es/pytorch_adapter.bin b/loras/ud/es/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1c7fb37108de76c68776a3d871f604bef70860d7 --- /dev/null +++ b/loras/ud/es/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:692c8f637078b0344cf043dc305d71f75f3b49c8e987e81daee2a1eb4442f15b +size 1333391 diff --git a/loras/ud/es/pytorch_model_head.bin b/loras/ud/es/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..0f72c0212f2d40689e52e9d8bc97a737b652d73c --- /dev/null +++ b/loras/ud/es/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a77a5b965dd85dc7913354ffa2cee69544c19976956e82b6b383ecee05a399e1 +size 342547 diff --git a/loras/ud/et/adapter_config.json b/loras/ud/et/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/et/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/et/head_config.json b/loras/ud/et/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/et/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/et/pytorch_adapter.bin b/loras/ud/et/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..945217f494dc946e486b436cde445fd51f6c6cd8 --- /dev/null +++ b/loras/ud/et/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e65f670d13b9c0514ce3254d1d486f9920602f96cffd59e1e8630f16a9546957 +size 1333391 diff --git a/loras/ud/et/pytorch_model_head.bin b/loras/ud/et/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..c8ffd54373b5bf66c810505c60577ca96bb7631a --- /dev/null +++ b/loras/ud/et/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c3e771b2b74d589d86918cfb48f336eaa897dab96ea01514c4c8ffe4689533c +size 342547 diff --git a/loras/ud/eu/adapter_config.json b/loras/ud/eu/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/eu/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/eu/head_config.json b/loras/ud/eu/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/eu/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/eu/pytorch_adapter.bin b/loras/ud/eu/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..de1c75800f895d60cb45ed1c0333bb152a08ff53 --- /dev/null +++ b/loras/ud/eu/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:141ec828a89690ef70bd23ea42c35e81d5baaa6a5c5a7370b1cfdd714d475fa6 +size 1333391 diff --git a/loras/ud/eu/pytorch_model_head.bin b/loras/ud/eu/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..70443bce9624eb3fe9700c71018fa132c3e91932 --- /dev/null +++ b/loras/ud/eu/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b362f9ff4c1f125d033f1f6f46c07f920e73ea9aa7c656f10af0866ed56cf2b +size 342547 diff --git a/loras/ud/fa/adapter_config.json b/loras/ud/fa/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/fa/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/fa/head_config.json b/loras/ud/fa/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/fa/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/fa/pytorch_adapter.bin b/loras/ud/fa/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..bd4d55570a88300b32953074b70ab56d4d20c8ad --- /dev/null +++ b/loras/ud/fa/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06ed92b5275bfbcf9932a1423eb4369a5851c72fd948ac75a2421d3a0d122602 +size 1333391 diff --git a/loras/ud/fa/pytorch_model_head.bin b/loras/ud/fa/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8fdae0e440e8cf3af52f9513c897fce2eb7db80c --- /dev/null +++ b/loras/ud/fa/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:668cf9a3e7f4f16a8730645fe9f58e670971eb05a913ecbd8b25538880e1f689 +size 342547 diff --git a/loras/ud/fi/adapter_config.json b/loras/ud/fi/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/fi/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/fi/head_config.json b/loras/ud/fi/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/fi/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/fi/pytorch_adapter.bin b/loras/ud/fi/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..13f55113b89089f415ab42a3928a61d99fd38e66 --- /dev/null +++ b/loras/ud/fi/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7cbdb84a77795688e4d926a670738a4fce7b1de4808bb33573fb0897292f4451 +size 1333391 diff --git a/loras/ud/fi/pytorch_model_head.bin b/loras/ud/fi/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..2c262859bc5c9bc0b7dbc00c2c27cd402cea534c --- /dev/null +++ b/loras/ud/fi/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8442d1b2527826d6314f5540f539e0724c2f1922e012937f21dbc5042bfc9325 +size 342547 diff --git a/loras/ud/fr/adapter_config.json b/loras/ud/fr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/fr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/fr/head_config.json b/loras/ud/fr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/fr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/fr/pytorch_adapter.bin b/loras/ud/fr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..347099fc597ddb6fadcc98809d56ad6732d8c1eb --- /dev/null +++ b/loras/ud/fr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf907b9f5e17d9130b482dc2b5ae54616a3659574965b040a0e326800d437d2f +size 1333391 diff --git a/loras/ud/fr/pytorch_model_head.bin b/loras/ud/fr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..4bac6e2d1b26efefc264ec600c008fabea63e58f --- /dev/null +++ b/loras/ud/fr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9056e2e9a619ee4098759f637203c1ac52264abef658b8971368e64427a3949 +size 342547 diff --git a/loras/ud/ga/adapter_config.json b/loras/ud/ga/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/ga/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ga/head_config.json b/loras/ud/ga/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/ga/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ga/pytorch_adapter.bin b/loras/ud/ga/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..cef264c4d3762f5915b7e317f6fefe62a2ebb0f7 --- /dev/null +++ b/loras/ud/ga/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec556b981be83cfd69e6cca5120a1d45e634013801fc32eeea4368545c8c1adb +size 1333391 diff --git a/loras/ud/ga/pytorch_model_head.bin b/loras/ud/ga/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..292015424bc9e6f2236aee38ba0bb6d4e70bb487 --- /dev/null +++ b/loras/ud/ga/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50adf0cb3050c4c18b0eb3c6ac271c24c8a35d12323cf520a83a3aeff6d46a13 +size 342547 diff --git a/loras/ud/gd/adapter_config.json b/loras/ud/gd/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/gd/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/gd/head_config.json b/loras/ud/gd/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/gd/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/gd/pytorch_adapter.bin b/loras/ud/gd/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..6fa689240d224c36f1ab7d0fd685fcf618a66943 --- /dev/null +++ b/loras/ud/gd/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:089e363b1dd62cfbd49838b93a37a575e426ba36d852a9925d8d9866d7cd0807 +size 1333391 diff --git a/loras/ud/gd/pytorch_model_head.bin b/loras/ud/gd/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..479356fd5fe6011996543c9a913f0d9e237bf2af --- /dev/null +++ b/loras/ud/gd/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba0d7df88fcef4b505a12bf5f4339ac1fb90637e593f17b56b6e49af27eac512 +size 342547 diff --git a/loras/ud/gl/adapter_config.json b/loras/ud/gl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/gl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/gl/head_config.json b/loras/ud/gl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/gl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/gl/pytorch_adapter.bin b/loras/ud/gl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..0b14958b64a5ffe1427d19127aaa487a697b8b98 --- /dev/null +++ b/loras/ud/gl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6469c5d34e3821f1bca26ce18d6164c999f2097b9ba66e578aef8c303f03d83 +size 1333391 diff --git a/loras/ud/gl/pytorch_model_head.bin b/loras/ud/gl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..564703cb54f3c422794926a3abf4f74e737b4023 --- /dev/null +++ b/loras/ud/gl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c08701aca41bf29fa28db15e9b73714b1b6d1cc324c85854d2a0f144d53f226a +size 342547 diff --git a/loras/ud/he/adapter_config.json b/loras/ud/he/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/he/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/he/head_config.json b/loras/ud/he/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/he/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/he/pytorch_adapter.bin b/loras/ud/he/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..0128986209898ad447c067b724114c169dec4a91 --- /dev/null +++ b/loras/ud/he/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9b47842ce46823046c560a5cfd2fd9072c972ecd92298ed5c3e0f1b98e28643 +size 1333391 diff --git a/loras/ud/he/pytorch_model_head.bin b/loras/ud/he/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..3c24ada3352403a3d012076300e516d0024ff08c --- /dev/null +++ b/loras/ud/he/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27704d510dc2efcc659fbc197f26a94f3293acb20b726609cc3337437f3b696e +size 342547 diff --git a/loras/ud/hi/adapter_config.json b/loras/ud/hi/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/hi/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/hi/head_config.json b/loras/ud/hi/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/hi/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/hi/pytorch_adapter.bin b/loras/ud/hi/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..50f1668361fa92fb720cf261397dcfd1cebd2b42 --- /dev/null +++ b/loras/ud/hi/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8ad97df1ab4c4c1ba9d9839801f59084a354c5a4bf03b4af76208d5c0145945 +size 1333391 diff --git a/loras/ud/hi/pytorch_model_head.bin b/loras/ud/hi/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..38f0a433a330deea6321482cf8d981ebb6a3f316 --- /dev/null +++ b/loras/ud/hi/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebae16348a5d9aadf7b21b5afd3548682161f7b54a40c678c9a839c502897a5d +size 342547 diff --git a/loras/ud/hu/adapter_config.json b/loras/ud/hu/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/hu/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/hu/head_config.json b/loras/ud/hu/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/hu/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/hu/pytorch_adapter.bin b/loras/ud/hu/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..a18ce40300f99cb587ff7d4cf75277eca5613db2 --- /dev/null +++ b/loras/ud/hu/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:035b3c19acf337fe9017968326305622d7d507a183361307d457c6e151a56bab +size 1333391 diff --git a/loras/ud/hu/pytorch_model_head.bin b/loras/ud/hu/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..c45df0bdbb24a7023172d84fc744cb1e990a308e --- /dev/null +++ b/loras/ud/hu/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:650c58717e765d6cd1aaf94ad8d7303b39fa2165477349a3dca9a89695d708bf +size 342547 diff --git a/loras/ud/hy/adapter_config.json b/loras/ud/hy/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/hy/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/hy/head_config.json b/loras/ud/hy/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/hy/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/hy/pytorch_adapter.bin b/loras/ud/hy/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..7a5c7f70cfb5d6f43ccc8a8cb882964495e13db1 --- /dev/null +++ b/loras/ud/hy/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2622c470ff56e6f5cc49d7c171a9b8f4938567d205eae9729d1279ce1f3b74f5 +size 1333391 diff --git a/loras/ud/hy/pytorch_model_head.bin b/loras/ud/hy/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..d573053ded45ed1caf65485a924bcae94878c139 --- /dev/null +++ b/loras/ud/hy/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38f9ea79cd4ec0456c4054295b1a85ecbcde9c59125392453e25bb9b23b6b4e4 +size 342547 diff --git a/loras/ud/is/adapter_config.json b/loras/ud/is/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/is/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/is/head_config.json b/loras/ud/is/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/is/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/is/pytorch_adapter.bin b/loras/ud/is/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..77d6dcd78a7930dd26ba7347690d2446818636aa --- /dev/null +++ b/loras/ud/is/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6641df9101b893ea27881492892b9c1cee8df5b37d45f8b5c914f1edf6a72be7 +size 1333391 diff --git a/loras/ud/is/pytorch_model_head.bin b/loras/ud/is/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..eeaa51b5083970cbebef25194e9d9b3c8ecd6d46 --- /dev/null +++ b/loras/ud/is/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ebeb54376dff44c397babdf8aeec78d998dc25081546ec74b4cb6044a9de1a8 +size 342547 diff --git a/loras/ud/it/adapter_config.json b/loras/ud/it/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/it/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/it/head_config.json b/loras/ud/it/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/it/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/it/pytorch_adapter.bin b/loras/ud/it/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..e4b41cb0ec7b78e0364172e9ec1e31f09577de37 --- /dev/null +++ b/loras/ud/it/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b984d292889a769ca2b11831a4c96e67eefda6da721408598c12243f3015ee8 +size 1333391 diff --git a/loras/ud/it/pytorch_model_head.bin b/loras/ud/it/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..b55fef44db4e7167c4004ae0efab428b755adb6e --- /dev/null +++ b/loras/ud/it/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91d2149525a84ff8407e4776518072d65ca6a20cf6a6ff1f1c3fb5a974d2c6a3 +size 342547 diff --git a/loras/ud/ja/adapter_config.json b/loras/ud/ja/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/ja/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ja/head_config.json b/loras/ud/ja/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/ja/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ja/pytorch_adapter.bin b/loras/ud/ja/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..b962b9bf6ecfed9bcec5c020304427f500d712ad --- /dev/null +++ b/loras/ud/ja/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ae98c555fdb0db7605ebb4b30a3334d9f5a6dfcd705239043e9022ba433539f +size 1333391 diff --git a/loras/ud/ja/pytorch_model_head.bin b/loras/ud/ja/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..0600b5736ca62193140a296687f2138d10c90559 --- /dev/null +++ b/loras/ud/ja/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d22de34b09d5892f025574ea1aeed8b050ee73e00bc79055cacc942d43a6c69 +size 342547 diff --git a/loras/ud/kk/adapter_config.json b/loras/ud/kk/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/kk/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/kk/head_config.json b/loras/ud/kk/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/kk/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/kk/pytorch_adapter.bin b/loras/ud/kk/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..cdd998ab887eb1cd746976e45ca587f53937fa2a --- /dev/null +++ b/loras/ud/kk/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:258841df879e4dff2d70dfdde81cb9967bfbd3dc937292b9dd630fc2644aecc0 +size 1333391 diff --git a/loras/ud/kk/pytorch_model_head.bin b/loras/ud/kk/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..fd98a0642fdaca7705aea0ba5af6a35cf571e80c --- /dev/null +++ b/loras/ud/kk/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90c782bad1134f112e8ffecd5d4667f21c04591b336d38fdf10bd8f22bf16bfd +size 342547 diff --git a/loras/ud/ko/adapter_config.json b/loras/ud/ko/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/ko/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ko/head_config.json b/loras/ud/ko/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/ko/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ko/pytorch_adapter.bin b/loras/ud/ko/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..36f56e5b543efd67c3820693fb26aaa31157f091 --- /dev/null +++ b/loras/ud/ko/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8b33d2b699be60716e9dab3b4f0f829b752c11d11ad139c01f637bbb6d7e08d +size 1333391 diff --git a/loras/ud/ko/pytorch_model_head.bin b/loras/ud/ko/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..3bccf27d428ec92631bbe4cf763fd39265d2079d --- /dev/null +++ b/loras/ud/ko/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a55a7fe8a600f33b8a144de3a7e46a7a517cc3f1a17db4ff4fdb09732c93cf62 +size 342547 diff --git a/loras/ud/la/adapter_config.json b/loras/ud/la/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/la/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/la/head_config.json b/loras/ud/la/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/la/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/la/pytorch_adapter.bin b/loras/ud/la/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..4957aa24a1bd4d4fdb96c53590a5edb82a443268 --- /dev/null +++ b/loras/ud/la/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c486c40242b1976d8e7e0f08635e79e88240bc91d149c9871a350a7cb733886e +size 1333391 diff --git a/loras/ud/la/pytorch_model_head.bin b/loras/ud/la/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8b686143d5ddd84bef7997d5f88f217df3dd6811 --- /dev/null +++ b/loras/ud/la/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5577156945a1f0c74d4cb18c3bc9d1ccfe36b0753f1b7fa5e75cef07505995fc +size 342547 diff --git a/loras/ud/lt/adapter_config.json b/loras/ud/lt/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/lt/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/lt/head_config.json b/loras/ud/lt/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/lt/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/lt/pytorch_adapter.bin b/loras/ud/lt/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..e087cbfcbab2423892b09e8bfdf3d8c1aaae7e8c --- /dev/null +++ b/loras/ud/lt/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4b319f396316df9ea47a9bae93194da3e463ad5379fce42eba1d5bcd75a6c1a +size 1333391 diff --git a/loras/ud/lt/pytorch_model_head.bin b/loras/ud/lt/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..8fe6a8bb4cb5b0860090d283785b7f484eb909be --- /dev/null +++ b/loras/ud/lt/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5b43e7482e2bca94bedfca562cb5c69777f803efb95dcc7a252c27591336fa7 +size 342547 diff --git a/loras/ud/lv/adapter_config.json b/loras/ud/lv/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/lv/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/lv/head_config.json b/loras/ud/lv/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/lv/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/lv/pytorch_adapter.bin b/loras/ud/lv/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..98c0e76f9e31c0a01f6739615ead5a261eec56b9 --- /dev/null +++ b/loras/ud/lv/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97d7dd08a7dd3895a61548327ec3eae96bc1948199d17a3bccc4adf0be1b7b0f +size 1333391 diff --git a/loras/ud/lv/pytorch_model_head.bin b/loras/ud/lv/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..a3076c0c4498815c90540e8df0e5864c5d1c5561 --- /dev/null +++ b/loras/ud/lv/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94cabc89c0e0eb6147c9ec78d7f21d25f6d4423626d7802dd70a9a76745e6cd9 +size 342547 diff --git a/loras/ud/mr/adapter_config.json b/loras/ud/mr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/mr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/mr/head_config.json b/loras/ud/mr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/mr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/mr/pytorch_adapter.bin b/loras/ud/mr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ed1f72d78a9136b530ca0079b37617045143ac20 --- /dev/null +++ b/loras/ud/mr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6f6fac54c0d467c349448d121df6bc5fc552a0379c2c7dc560575d33d453878 +size 1333391 diff --git a/loras/ud/mr/pytorch_model_head.bin b/loras/ud/mr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..3931f00c17b557713fa31d4077a2a157606ee6a4 --- /dev/null +++ b/loras/ud/mr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4bf26583b0e664875b361baccbd4ac62ebba42719d7b5648576f9fc5d56d4a7b +size 342547 diff --git a/loras/ud/mt/adapter_config.json b/loras/ud/mt/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/mt/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/mt/head_config.json b/loras/ud/mt/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/mt/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/mt/pytorch_adapter.bin b/loras/ud/mt/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..236a95d4d61d3b56dabeb9d1a47945ff57b7b6e7 --- /dev/null +++ b/loras/ud/mt/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f57904c270b2a6c2b020dd8c07406da4a6cd43587db2722c92010dc3bdd0162 +size 1333391 diff --git a/loras/ud/mt/pytorch_model_head.bin b/loras/ud/mt/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..14a64e6c7333a6e3392e5ac24c40e72dc0f5d9a3 --- /dev/null +++ b/loras/ud/mt/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6973f378f1c3ee0c6e5a30207509e3f3b866c223fd5d13840c59c74f44708fc7 +size 342547 diff --git a/loras/ud/nl/adapter_config.json b/loras/ud/nl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/nl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/nl/head_config.json b/loras/ud/nl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/nl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/nl/pytorch_adapter.bin b/loras/ud/nl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..e02d3d510105c4037d63f82d690939c5ec0fddf6 --- /dev/null +++ b/loras/ud/nl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:505a21080a741e59c49c77f3e6b542ddf6a9228c97ca48315c4bfaea07d156fb +size 1333391 diff --git a/loras/ud/nl/pytorch_model_head.bin b/loras/ud/nl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..9ef60ae8503f06f6ae51499fe701b86a12656c04 --- /dev/null +++ b/loras/ud/nl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9fade1c10f9392f5ab3a49a0e0a8f9602905641498a4a698826e1ebdfdfddd84 +size 342547 diff --git a/loras/ud/no/adapter_config.json b/loras/ud/no/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/no/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/no/head_config.json b/loras/ud/no/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/no/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/no/pytorch_adapter.bin b/loras/ud/no/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..b248071ef62087f5de428f6c4500c583705ec535 --- /dev/null +++ b/loras/ud/no/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5654f7fda309f6aba177a992d666dd555f2f822b58ca7d5d70274ea333f7e1a9 +size 1333391 diff --git a/loras/ud/no/pytorch_model_head.bin b/loras/ud/no/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..93ad1dfcbcf5fc9b35987f2e92066eb0fcc5c33a --- /dev/null +++ b/loras/ud/no/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c899299e5f738eb15b30691f4b30b512fadf00eb50b113790932bf669c39344 +size 342547 diff --git a/loras/ud/pl/adapter_config.json b/loras/ud/pl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/pl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/pl/head_config.json b/loras/ud/pl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/pl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/pl/pytorch_adapter.bin b/loras/ud/pl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..d2ee1c8779de177025ed972277a9f9cc10b36402 --- /dev/null +++ b/loras/ud/pl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2706426fdc6fedb4b9a2f6c8271195ffb05ecd55e4d396bec2e5e5f0a286349a +size 1333391 diff --git a/loras/ud/pl/pytorch_model_head.bin b/loras/ud/pl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..6e773887d3b3629b026507bc53c8f9ca93a53010 --- /dev/null +++ b/loras/ud/pl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93692335cc67488ce4d952c97ecb91e1facc1e4ce137850696c4c6e2d04b849e +size 342547 diff --git a/loras/ud/pt/adapter_config.json b/loras/ud/pt/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/pt/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/pt/head_config.json b/loras/ud/pt/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/pt/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/pt/pytorch_adapter.bin b/loras/ud/pt/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..0cdafdf958f23385aeb122e94e7b0c4371bde1c1 --- /dev/null +++ b/loras/ud/pt/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57986cd692a8a935be22eb7696d53964cff73dce13a5401cdc3bbd5c9ab3ff40 +size 1333391 diff --git a/loras/ud/pt/pytorch_model_head.bin b/loras/ud/pt/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..cb09841d3150f5e3cb907d7d9cb3984190ea76c4 --- /dev/null +++ b/loras/ud/pt/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e862236d3282cc1cfa4df8c4acb9472e2e3ef1683a568a64479adcc076424e6d +size 342547 diff --git a/loras/ud/ro/adapter_config.json b/loras/ud/ro/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/ro/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ro/head_config.json b/loras/ud/ro/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/ro/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ro/pytorch_adapter.bin b/loras/ud/ro/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ebbd1be69343e1a6c44c3e90be842210adf7e0b --- /dev/null +++ b/loras/ud/ro/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6980099ccae016c315cd2d1b335b9bd299e3b326635be9da98ca9b434c7644dc +size 1333391 diff --git a/loras/ud/ro/pytorch_model_head.bin b/loras/ud/ro/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..13cb13ffb3fdbb861ba2e483a9db4cb00717203f --- /dev/null +++ b/loras/ud/ro/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae38423cd9c3184cc77bd90a345d1ec8633377f861f1c59472e75298c387b228 +size 342547 diff --git a/loras/ud/ru/adapter_config.json b/loras/ud/ru/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/ru/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ru/head_config.json b/loras/ud/ru/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/ru/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ru/pytorch_adapter.bin b/loras/ud/ru/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..ffe32cd62b1da46f7415cf3e4df07fe7c9742597 --- /dev/null +++ b/loras/ud/ru/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ac4253dedf4b42953e208e28f43a8cd62e7cd49fed4fee5aec1e5099e03b1b0 +size 1333391 diff --git a/loras/ud/ru/pytorch_model_head.bin b/loras/ud/ru/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..dff09a435775c746fd2b6116fd38f0c9db24d9a5 --- /dev/null +++ b/loras/ud/ru/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa4bded65ed2cd66686b7db0ba87fcc8b53eb3bd9b8514d09611e7c873eff54a +size 342547 diff --git a/loras/ud/sk/adapter_config.json b/loras/ud/sk/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/sk/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/sk/head_config.json b/loras/ud/sk/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/sk/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/sk/pytorch_adapter.bin b/loras/ud/sk/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..52026cc24789d9b84c3dfb597ccb9da9466fd940 --- /dev/null +++ b/loras/ud/sk/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb9c85da72be91b2dadacfb49e0e8d0281a5f2ecf74f70f58c0a00bb6aff077f +size 1333391 diff --git a/loras/ud/sk/pytorch_model_head.bin b/loras/ud/sk/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..d3435b7e160f2a32cb691c7b1c7b6bebc98f8c27 --- /dev/null +++ b/loras/ud/sk/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8546548fba2989209a8f0d2105eae294997b96b827f55155eef88cbadb554fcb +size 342547 diff --git a/loras/ud/sl/adapter_config.json b/loras/ud/sl/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/sl/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/sl/head_config.json b/loras/ud/sl/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/sl/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/sl/pytorch_adapter.bin b/loras/ud/sl/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..c7ccba0be28593aa5bf0609c6c28be09fc8f2dde --- /dev/null +++ b/loras/ud/sl/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d42184cbbfed1ee58d50a55ad681dd25c893edff2e5408b3d88722f08a11c87 +size 1333391 diff --git a/loras/ud/sl/pytorch_model_head.bin b/loras/ud/sl/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..3c81654e72d41acc4de8de927a201b16ad73dc9c --- /dev/null +++ b/loras/ud/sl/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a561cdb2aedd751e6caff4f28a019ecca329ae8054a41391dbf8b28469db433d +size 342547 diff --git a/loras/ud/sr/adapter_config.json b/loras/ud/sr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/sr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/sr/head_config.json b/loras/ud/sr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/sr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/sr/pytorch_adapter.bin b/loras/ud/sr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..6cb52055141106a855dbe384c4329ff4824ca1fc --- /dev/null +++ b/loras/ud/sr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d63313e60c9f5188324cb9458f076cadd55fb87b0a53c1feed41606988b549f1 +size 1333391 diff --git a/loras/ud/sr/pytorch_model_head.bin b/loras/ud/sr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..7f3feb3ed35d036c858e71ca793c97ef66e8de14 --- /dev/null +++ b/loras/ud/sr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9282a00f0592124c07d3ce7e7ee9660181ddfec27715d677ff6859cd00460389 +size 342547 diff --git a/loras/ud/sv/adapter_config.json b/loras/ud/sv/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/sv/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/sv/head_config.json b/loras/ud/sv/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/sv/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/sv/pytorch_adapter.bin b/loras/ud/sv/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..1aae93704f4036d6929cc42d7394d2f68ecd5253 --- /dev/null +++ b/loras/ud/sv/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e39360cb352ac7b6c5507a9049bf9d5b3941721b0845c56feaa653ffacf86e6 +size 1333391 diff --git a/loras/ud/sv/pytorch_model_head.bin b/loras/ud/sv/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..769fe748a1827056b7a76bbab517aa4798c0876c --- /dev/null +++ b/loras/ud/sv/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e071731c45972f3388d1164e74b9945ecbc977e484d316f25c2bcfa61fe5da2f +size 342547 diff --git a/loras/ud/ta/adapter_config.json b/loras/ud/ta/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/ta/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ta/head_config.json b/loras/ud/ta/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/ta/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ta/pytorch_adapter.bin b/loras/ud/ta/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..a8611cf4c3b81b04fe69e247b4463b06bbff56a5 --- /dev/null +++ b/loras/ud/ta/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9a42132c9d0bafd8af67da6ea9fa1a8566714abede5485c1153519cb3346c9f +size 1333391 diff --git a/loras/ud/ta/pytorch_model_head.bin b/loras/ud/ta/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..2077ad24b379913059b1d6aab6ff7c3abbb72983 --- /dev/null +++ b/loras/ud/ta/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0dc35f89acbcb36673c9bf067090d5bafcd8471a4dbffba9f02503a6bb1fd786 +size 342547 diff --git a/loras/ud/tr/adapter_config.json b/loras/ud/tr/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/tr/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/tr/head_config.json b/loras/ud/tr/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/tr/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/tr/pytorch_adapter.bin b/loras/ud/tr/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..b9f8b7e29d354751cb9c4f4c3cd7bf192dbfb64f --- /dev/null +++ b/loras/ud/tr/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86b59ad09723009872360c289d3634e18a488b1b36d1959746b2b7014b2eb9e0 +size 1333391 diff --git a/loras/ud/tr/pytorch_model_head.bin b/loras/ud/tr/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..f1ccb872fcd68738700141705efd9cb860f7bf1f --- /dev/null +++ b/loras/ud/tr/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ddb7a5cf3e1ca395f110cb3e092839ebdba481778f2f2d76db4227868ac2ecc +size 342547 diff --git a/loras/ud/uk/adapter_config.json b/loras/ud/uk/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/uk/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/uk/head_config.json b/loras/ud/uk/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/uk/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/uk/pytorch_adapter.bin b/loras/ud/uk/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..a0982fde28d6846dfd47c5d8a48f8cd9e0efb249 --- /dev/null +++ b/loras/ud/uk/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6455d17fa717456f5b89cd75ae18ce7b75b1b491a8c87704fc0a7329d944d751 +size 1333391 diff --git a/loras/ud/uk/pytorch_model_head.bin b/loras/ud/uk/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..03c3ab5d075d294bd85d68c38da7bf08d028d615 --- /dev/null +++ b/loras/ud/uk/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d95634b159a122a2f0f8f0b72528e86ae00dad07e50357143be1040bbb71a157 +size 342547 diff --git a/loras/ud/ur/adapter_config.json b/loras/ud/ur/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/ur/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ur/head_config.json b/loras/ud/ur/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/ur/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/ur/pytorch_adapter.bin b/loras/ud/ur/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..02c7dc29b830d59af703987fef774e562aff2f3b --- /dev/null +++ b/loras/ud/ur/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:884d5c7bea14ba5429d531e14bcccfdb4308da7029ee8a865c2b0c58eb2f7745 +size 1333391 diff --git a/loras/ud/ur/pytorch_model_head.bin b/loras/ud/ur/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..530ac9024dfe35c1b2c4b250229334634d19b861 --- /dev/null +++ b/loras/ud/ur/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81e6eae34710d036432a8d14383cda35119216a4bcb6b44c52b0d8313b76f725 +size 342547 diff --git a/loras/ud/vi/adapter_config.json b/loras/ud/vi/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/vi/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/vi/head_config.json b/loras/ud/vi/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/vi/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/vi/pytorch_adapter.bin b/loras/ud/vi/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..715c34a4f187965685a1b9d32e943cc246e1143d --- /dev/null +++ b/loras/ud/vi/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3dd5dc02d975b9e340b03052ff9b908a84fe9495113c7bc811d736e8af273a1 +size 1333391 diff --git a/loras/ud/vi/pytorch_model_head.bin b/loras/ud/vi/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..d4389ba2a4c5dc6c0a1e7308f14f1e65eeb03610 --- /dev/null +++ b/loras/ud/vi/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c46893b5b7e303d2275b8ef3767085355e8051e4d0fa50b3eed03cab6030fe6c +size 342547 diff --git a/loras/ud/zh/adapter_config.json b/loras/ud/zh/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..223a42003e70eaf3fb1b9a5344aa3f54bba35bee --- /dev/null +++ b/loras/ud/zh/adapter_config.json @@ -0,0 +1,25 @@ +{ + "config": { + "alpha": 32, + "architecture": "lora", + "attn_matrices": [ + "q", + "v" + ], + "composition_mode": "add", + "dropout": 0.0, + "init_weights": "lora", + "intermediate_lora": true, + "leave_out": [], + "output_lora": false, + "r": 16, + "selfattn_lora": true, + "use_gating": false + }, + "hidden_size": 768, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": "text", + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/zh/head_config.json b/loras/ud/zh/head_config.json new file mode 100644 index 0000000000000000000000000000000000000000..e8fc67a978a1781a25687b0992bc7acc1ab88fb5 --- /dev/null +++ b/loras/ud/zh/head_config.json @@ -0,0 +1,123 @@ +{ + "config": null, + "hidden_size": 768, + "label2id": { + "LABEL_0": 0, + "LABEL_1": 1, + "LABEL_10": 10, + "LABEL_100": 100, + "LABEL_101": 101, + "LABEL_102": 102, + "LABEL_103": 103, + "LABEL_104": 104, + "LABEL_105": 105, + "LABEL_106": 106, + "LABEL_107": 107, + "LABEL_108": 108, + "LABEL_109": 109, + "LABEL_11": 11, + "LABEL_110": 110, + "LABEL_12": 12, + "LABEL_13": 13, + "LABEL_14": 14, + "LABEL_15": 15, + "LABEL_16": 16, + "LABEL_17": 17, + "LABEL_18": 18, + "LABEL_19": 19, + "LABEL_2": 2, + "LABEL_20": 20, + "LABEL_21": 21, + "LABEL_22": 22, + "LABEL_23": 23, + "LABEL_24": 24, + "LABEL_25": 25, + "LABEL_26": 26, + "LABEL_27": 27, + "LABEL_28": 28, + "LABEL_29": 29, + "LABEL_3": 3, + "LABEL_30": 30, + "LABEL_31": 31, + "LABEL_32": 32, + "LABEL_33": 33, + "LABEL_34": 34, + "LABEL_35": 35, + "LABEL_36": 36, + "LABEL_37": 37, + "LABEL_38": 38, + "LABEL_39": 39, + "LABEL_4": 4, + "LABEL_40": 40, + "LABEL_41": 41, + "LABEL_42": 42, + "LABEL_43": 43, + "LABEL_44": 44, + "LABEL_45": 45, + "LABEL_46": 46, + "LABEL_47": 47, + "LABEL_48": 48, + "LABEL_49": 49, + "LABEL_5": 5, + "LABEL_50": 50, + "LABEL_51": 51, + "LABEL_52": 52, + "LABEL_53": 53, + "LABEL_54": 54, + "LABEL_55": 55, + "LABEL_56": 56, + "LABEL_57": 57, + "LABEL_58": 58, + "LABEL_59": 59, + "LABEL_6": 6, + "LABEL_60": 60, + "LABEL_61": 61, + "LABEL_62": 62, + "LABEL_63": 63, + "LABEL_64": 64, + "LABEL_65": 65, + "LABEL_66": 66, + "LABEL_67": 67, + "LABEL_68": 68, + "LABEL_69": 69, + "LABEL_7": 7, + "LABEL_70": 70, + "LABEL_71": 71, + "LABEL_72": 72, + "LABEL_73": 73, + "LABEL_74": 74, + "LABEL_75": 75, + "LABEL_76": 76, + "LABEL_77": 77, + "LABEL_78": 78, + "LABEL_79": 79, + "LABEL_8": 8, + "LABEL_80": 80, + "LABEL_81": 81, + "LABEL_82": 82, + "LABEL_83": 83, + "LABEL_84": 84, + "LABEL_85": 85, + "LABEL_86": 86, + "LABEL_87": 87, + "LABEL_88": 88, + "LABEL_89": 89, + "LABEL_9": 9, + "LABEL_90": 90, + "LABEL_91": 91, + "LABEL_92": 92, + "LABEL_93": 93, + "LABEL_94": 94, + "LABEL_95": 95, + "LABEL_96": 96, + "LABEL_97": 97, + "LABEL_98": 98, + "LABEL_99": 99 + }, + "model_class": "SubwordXLMForTokenClassification", + "model_name": "xlmr-3l-v3_look48_lc0.1-mix2", + "model_type": "xlm-token", + "name": null, + "num_labels": 111, + "version": "0.1.1" +} \ No newline at end of file diff --git a/loras/ud/zh/pytorch_adapter.bin b/loras/ud/zh/pytorch_adapter.bin new file mode 100644 index 0000000000000000000000000000000000000000..57330e84afaef86f0157ad0c7f240bed5bd4f226 --- /dev/null +++ b/loras/ud/zh/pytorch_adapter.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3bab67da284b0bf4d378f5991587a744c98d476b65aee98fbea0bb450464d480 +size 1333391 diff --git a/loras/ud/zh/pytorch_model_head.bin b/loras/ud/zh/pytorch_model_head.bin new file mode 100644 index 0000000000000000000000000000000000000000..25cba4277834b2619cc4e4d43807418977e6ab22 --- /dev/null +++ b/loras/ud/zh/pytorch_model_head.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2c29c32233805a4c422503ef5751e58a65bcd0fed6d40d7bc17537435a7f1e1 +size 342547