Saving weights of epoch 2 at step 74
Browse files
flax_model.msgpack
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5262347427
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f62dbb35840c3ee0b6e92eefbf42206f23b69c4d9bb70730914a09033cf3313b
|
3 |
size 5262347427
|
results_tensorboard/{events.out.tfevents.1627575327.t1v-n-cb40e504-w-0.12731.3.v2 → events.out.tfevents.1627575762.t1v-n-cb40e504-w-0.14227.3.v2}
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:589f97a8921bac4055e7af9db80f84ebc5d41840ce4ffd6839b04da359c9de3e
|
3 |
+
size 13044
|
src/gptneo_story.py
CHANGED
@@ -44,8 +44,8 @@ def preprocess(example):
|
|
44 |
train_dataset=train_dataset.map(preprocess)
|
45 |
validation_dataset=validation_dataset.map(preprocess)
|
46 |
|
47 |
-
len_train_dataset=
|
48 |
-
len_validation_dataset=
|
49 |
|
50 |
train_dataset=train_dataset.select(range(len_train_dataset))
|
51 |
validation_dataset=validation_dataset.select(range(len_validation_dataset))
|
@@ -53,7 +53,7 @@ validation_dataset=validation_dataset.select(range(len_validation_dataset))
|
|
53 |
remove_col=train_dataset.column_names
|
54 |
|
55 |
def tokenize(examples):
|
56 |
-
tokenized_examples=tokenizer(examples['first_sentence'],examples['second_sentence'],padding='max_length',truncation=True,max_length=
|
57 |
tokenized_examples['labels']=int(examples['right_answer'])
|
58 |
return tokenized_examples
|
59 |
|
|
|
44 |
train_dataset=train_dataset.map(preprocess)
|
45 |
validation_dataset=validation_dataset.map(preprocess)
|
46 |
|
47 |
+
len_train_dataset=2456
|
48 |
+
len_validation_dataset=1200
|
49 |
|
50 |
train_dataset=train_dataset.select(range(len_train_dataset))
|
51 |
validation_dataset=validation_dataset.select(range(len_validation_dataset))
|
|
|
53 |
remove_col=train_dataset.column_names
|
54 |
|
55 |
def tokenize(examples):
|
56 |
+
tokenized_examples=tokenizer(examples['first_sentence'],examples['second_sentence'],padding='max_length',truncation=True,max_length=339,return_tensors='jax')
|
57 |
tokenized_examples['labels']=int(examples['right_answer'])
|
58 |
return tokenized_examples
|
59 |
|