medhabi commited on
Commit
a756e26
1 Parent(s): 925ecf0
checkponts/checkpont-1/context.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec6e8a7087d30885eda868195e8fcfe392e2211c169e035b2dcd9d70e1462dc3
3
+ size 431
checkponts/checkpont-1/lr_scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d42493a1084eba4421ce623c0550a88535c1a058c3497cd64faf57245b07e5c
3
+ size 559
checkponts/checkpont-1/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f16ac4aaf4ff9ebe77d1ac89ba7af938be9453b0ae5568bcbe94ff29dcaf1c8
3
+ size 265506677
checkponts/checkpont-1/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d95c520e4f7b3d9f32c55a2b4d3c744d06eff18557e49e8bf1f7cd7a9b1983de
3
+ size 530992489
checkponts/checkpont-2/context.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8316cc864404584f9e561aa71c5c359f12e869fc7ee9992b1532cfc05525c242
3
+ size 431
checkponts/checkpont-2/lr_scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf789cc78c162eba88c2cdbbf2870ce781be111bf50fe8d36ebe79eab5067c9f
3
+ size 559
checkponts/checkpont-2/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1add215cda5584ca44f5b772051a7719d2946a882a81ace3dcf9bac7f2d41496
3
+ size 265506677
checkponts/checkpont-2/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d540d53043b316562de0306cfbd7e57eee5314f81ffec282950a266acd0e9ea0
3
+ size 530992489
checkponts/checkpont-3/context.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e987eb690d97093447db9db9ab6bb427cb66344b4f0978646e79ffa98ea06ac
3
+ size 431
checkponts/checkpont-3/lr_scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a80afcf075552cc666b3001775e76849cefa3bc3fd53aa76b084a7b94f95ccd2
3
+ size 559
checkponts/checkpont-3/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87c2f7fb51b49024b067a82b142b098fc366a6e632fac5cac0b5a5cdb0129212
3
+ size 265506677
checkponts/checkpont-3/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:675d0869dcabcaa7521c045bdd4fe78c10b5da3e378245b08887e2b658082389
3
+ size 530992489
checkponts/checkpont-best/context.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec6e8a7087d30885eda868195e8fcfe392e2211c169e035b2dcd9d70e1462dc3
3
+ size 431
checkponts/checkpont-best/lr_scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d42493a1084eba4421ce623c0550a88535c1a058c3497cd64faf57245b07e5c
3
+ size 559
checkponts/checkpont-best/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f16ac4aaf4ff9ebe77d1ac89ba7af938be9453b0ae5568bcbe94ff29dcaf1c8
3
+ size 265506677
checkponts/checkpont-best/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d95c520e4f7b3d9f32c55a2b4d3c744d06eff18557e49e8bf1f7cd7a9b1983de
3
+ size 530992489
config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "architectures": [
3
  "BERTTextToRating"
4
  ],
 
1
  {
2
+ "_name_or_path": "distilbert-base-uncased-score-pred",
3
  "architectures": [
4
  "BERTTextToRating"
5
  ],
info.txt ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ AdamW(model.parameters(), lr=5e-5)
2
+
3
+ class BERTTextToRating(PreTrainedModel):
4
+ config_class = BERTTextToRatingConfig
5
+
6
+ def __init__(self, config):
7
+ super(BERTTextToRating, self).__init__(config)
8
+ model_checkpoint = "medhabi/distilbert-base-uncased-mlm-ta-local"
9
+ model = AutoModelForMaskedLM.from_pretrained(model_checkpoint)
10
+ self.bert_model = model.distilbert
11
+ self.dropout = torch.nn.Dropout(0.3)
12
+ self.linear = torch.nn.Linear(768, 5)
13
+
14
+ def forward(self, input_ids, attention_mask, token_type_ids):
15
+ output = self.bert_model(
16
+ input_ids,
17
+ attention_mask=attention_mask,
18
+ )
19
+ output_dropout = self.dropout(output.last_hidden_state[0][0].reshape(1,-1))
20
+ output = self.linear(output_dropout)
21
+ return output
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:023b147bf59c5007a31d367591f5c24980200191dc85346342c8b4a96fe41e5a
3
- size 265506677
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a48212f4032484ca71942cffd300f7afc2de2825377533850036133c7bfd2b1d
3
+ size 265505269