ckandemir commited on
Commit
5f81f8e
1 Parent(s): 046825e

Training completed!

Browse files
Files changed (4) hide show
  1. README.md +20 -22
  2. config.json +6 -2
  3. pytorch_model.bin +2 -2
  4. training_args.bin +1 -1
README.md CHANGED
@@ -3,39 +3,34 @@ license: mit
3
  base_model: xlm-roberta-base
4
  tags:
5
  - generated_from_trainer
6
- - NER
7
- - crypto
8
  metrics:
9
  - f1
10
  model-index:
11
- - name: xlm-roberta-base-finetuned-ner-crypto
12
  results: []
13
- widget:
14
- - text: "Didn't I tell you that that was a decent entry point on $PROPHET? If you are in - congrats, Prophet is up 90% in the last 2 weeks and 50% up in the last week alone"
15
-
16
  ---
17
 
18
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
19
  should probably proofread and complete it, then remove this comment. -->
20
 
21
- # xlm-roberta-base-finetuned-ner-crypto
22
 
23
  This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the None dataset.
24
  It achieves the following results on the evaluation set:
25
- - Loss: 0.0100
26
- - F1: 0.9883
27
 
28
  ## Model description
29
- This model is a fine-tuned version of xlm-roberta-base, specializing in Named Entity Recognition (NER) within the cryptocurrency domain. It is optimized to recognize and classify entities such as cryptocurrency ticker symbols, names, and addresses within text.
30
 
31
- ## Intended uses
32
- Designed primarily for NER tasks in the cryptocurrency sector, this model excels in identifying and categorizing ticker symbols, cryptocurrency names, and addresses in textual content.
33
- ## Limitations
34
- The model might not perform well in identifying and classifying entities that were not part of the training data or those that are less frequent in the cryptocurrency domain. It may also be sensitive to the context and format in which the entities are presented.
 
35
 
36
  ## Training and evaluation data
37
 
38
- The model was trained using a diverse dataset, including artificially generated tweets and ERC20 token metadata fetched through the [Covalent API](https://www.covalenthq.com/docs/unified-api/). GPT was employed to generate 500 synthetic tweets tailored for the cryptocurrency domain. The Covalent API was instrumental in obtaining a rich set of unique ERC20 token metadata entries, enhancing the model's understanding and recognition of cryptocurrency entities.
39
 
40
  ## Training procedure
41
 
@@ -43,25 +38,28 @@ The model was trained using a diverse dataset, including artificially generated
43
 
44
  The following hyperparameters were used during training:
45
  - learning_rate: 5e-05
46
- - train_batch_size: 24
47
- - eval_batch_size: 24
48
  - seed: 42
49
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: linear
51
- - num_epochs: 3
52
 
53
  ### Training results
54
 
55
  | Training Loss | Epoch | Step | Validation Loss | F1 |
56
  |:-------------:|:-----:|:----:|:---------------:|:------:|
57
- | 0.0338 | 1.0 | 1000 | 0.0126 | 0.9877 |
58
- | 0.0105 | 2.0 | 2000 | 0.0112 | 0.9867 |
59
- | 0.0081 | 3.0 | 3000 | 0.0100 | 0.9883 |
 
 
 
60
 
61
 
62
  ### Framework versions
63
 
64
  - Transformers 4.34.1
65
  - Pytorch 2.1.0+cu118
66
- - Datasets 2.14.5
67
  - Tokenizers 0.14.1
 
3
  base_model: xlm-roberta-base
4
  tags:
5
  - generated_from_trainer
 
 
6
  metrics:
7
  - f1
8
  model-index:
9
+ - name: xlm-roberta-base-finetuned-NER-crypto
10
  results: []
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
+ # xlm-roberta-base-finetuned-NER-crypto
17
 
18
  This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.0041
21
+ - F1: 0.9960
22
 
23
  ## Model description
 
24
 
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
 
31
  ## Training and evaluation data
32
 
33
+ More information needed
34
 
35
  ## Training procedure
36
 
 
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 5e-05
41
+ - train_batch_size: 32
42
+ - eval_batch_size: 32
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
+ - num_epochs: 6
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | F1 |
51
  |:-------------:|:-----:|:----:|:---------------:|:------:|
52
+ | 0.1208 | 1.0 | 125 | 0.0181 | 0.9872 |
53
+ | 0.0061 | 2.0 | 250 | 0.0055 | 0.9951 |
54
+ | 0.0028 | 3.0 | 375 | 0.0037 | 0.9948 |
55
+ | 0.002 | 4.0 | 500 | 0.0037 | 0.9960 |
56
+ | 0.0016 | 5.0 | 625 | 0.0040 | 0.9960 |
57
+ | 0.0013 | 6.0 | 750 | 0.0041 | 0.9960 |
58
 
59
 
60
  ### Framework versions
61
 
62
  - Transformers 4.34.1
63
  - Pytorch 2.1.0+cu118
64
+ - Datasets 2.14.6
65
  - Tokenizers 0.14.1
config.json CHANGED
@@ -17,18 +17,22 @@
17
  "3": "I-NAME",
18
  "4": "B-TICKER_SYMBOL",
19
  "5": "I-TICKER_SYMBOL",
20
- "6": "O"
 
 
21
  },
22
  "initializer_range": 0.02,
23
  "intermediate_size": 3072,
24
  "label2id": {
25
  "B-ADDRESS": 0,
 
26
  "B-NAME": 2,
27
  "B-TICKER_SYMBOL": 4,
28
  "I-ADDRESS": 1,
 
29
  "I-NAME": 3,
30
  "I-TICKER_SYMBOL": 5,
31
- "O": 6
32
  },
33
  "layer_norm_eps": 1e-05,
34
  "max_position_embeddings": 514,
 
17
  "3": "I-NAME",
18
  "4": "B-TICKER_SYMBOL",
19
  "5": "I-TICKER_SYMBOL",
20
+ "6": "B-CHAIN",
21
+ "7": "I-CHAIN",
22
+ "8": "O"
23
  },
24
  "initializer_range": 0.02,
25
  "intermediate_size": 3072,
26
  "label2id": {
27
  "B-ADDRESS": 0,
28
+ "B-CHAIN": 6,
29
  "B-NAME": 2,
30
  "B-TICKER_SYMBOL": 4,
31
  "I-ADDRESS": 1,
32
+ "I-CHAIN": 7,
33
  "I-NAME": 3,
34
  "I-TICKER_SYMBOL": 5,
35
+ "O": 8
36
  },
37
  "layer_norm_eps": 1e-05,
38
  "max_position_embeddings": 514,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f0a69c1efc354106cedf2ffd3ea3c8047fc0ee94a1db45fa0625c0e880969bf
3
- size 1109902502
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db46f3eedb6f759c61837663da1c8a0b1ef5f93a3b719236a354f995b5ccc11b
3
+ size 1109908646
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dcd4d8e12a0b74cf2d0e3c758b224bc40e3b0dc0b86c2e36ff5abfc32839ad37
3
  size 4536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:473a9325e049fc3b2929c677b29a23cb1de27175e31d340329803c03f19a81ab
3
  size 4536