Vivek commited on
Commit
89ba378
1 Parent(s): 2b36275

Saving weights of epoch 2 at step 74

Browse files
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:026a9da1385727ad25b118be45961965cf77faaf56bfa6193d1b1105fd73fc2c
3
  size 5262347427
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f62dbb35840c3ee0b6e92eefbf42206f23b69c4d9bb70730914a09033cf3313b
3
  size 5262347427
results_tensorboard/{events.out.tfevents.1627575327.t1v-n-cb40e504-w-0.12731.3.v2 → events.out.tfevents.1627575762.t1v-n-cb40e504-w-0.14227.3.v2} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8da59be43f840d393332e6be49c905a9a5e49e95cfb3aa2d172593aa454e3283
3
- size 1160
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:589f97a8921bac4055e7af9db80f84ebc5d41840ce4ffd6839b04da359c9de3e
3
+ size 13044
src/gptneo_story.py CHANGED
@@ -44,8 +44,8 @@ def preprocess(example):
44
  train_dataset=train_dataset.map(preprocess)
45
  validation_dataset=validation_dataset.map(preprocess)
46
 
47
- len_train_dataset=100
48
- len_validation_dataset=100
49
 
50
  train_dataset=train_dataset.select(range(len_train_dataset))
51
  validation_dataset=validation_dataset.select(range(len_validation_dataset))
@@ -53,7 +53,7 @@ validation_dataset=validation_dataset.select(range(len_validation_dataset))
53
  remove_col=train_dataset.column_names
54
 
55
  def tokenize(examples):
56
- tokenized_examples=tokenizer(examples['first_sentence'],examples['second_sentence'],padding='max_length',truncation=True,max_length=256,return_tensors='jax')
57
  tokenized_examples['labels']=int(examples['right_answer'])
58
  return tokenized_examples
59
 
44
  train_dataset=train_dataset.map(preprocess)
45
  validation_dataset=validation_dataset.map(preprocess)
46
 
47
+ len_train_dataset=2456
48
+ len_validation_dataset=1200
49
 
50
  train_dataset=train_dataset.select(range(len_train_dataset))
51
  validation_dataset=validation_dataset.select(range(len_validation_dataset))
53
  remove_col=train_dataset.column_names
54
 
55
  def tokenize(examples):
56
+ tokenized_examples=tokenizer(examples['first_sentence'],examples['second_sentence'],padding='max_length',truncation=True,max_length=339,return_tensors='jax')
57
  tokenized_examples['labels']=int(examples['right_answer'])
58
  return tokenized_examples
59