pere commited on
Commit
9e311e8
·
1 Parent(s): 9f37347

Saving weights and logs of step 10001

Browse files
config.json CHANGED
@@ -21,7 +21,7 @@
21
  "pad_token_id": 1,
22
  "position_embedding_type": "absolute",
23
  "torch_dtype": "float32",
24
- "transformers_version": "4.25.1",
25
  "type_vocab_size": 1,
26
  "use_cache": true,
27
  "vocab_size": 250002
 
21
  "pad_token_id": 1,
22
  "position_embedding_type": "absolute",
23
  "torch_dtype": "float32",
24
+ "transformers_version": "4.26.0",
25
  "type_vocab_size": 1,
26
  "use_cache": true,
27
  "vocab_size": 250002
events.out.tfevents.1674763276.t1v-n-d7824f3a-w-7.25607.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fce5716dd72aa0242342a6d8b296bb75b4f03956d9e3832d91bd953ff425c56
3
+ size 1471447
flax_model.msgpack ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f38cecbf3ccd025280b07cdc9588400e58dbd0c90d4de5bb4c28d7eb466b67b5
3
+ size 1113187999
run_mlm_flax_stream.py CHANGED
@@ -564,7 +564,7 @@ if __name__ == "__main__":
564
  train_metrics = []
565
  eval_metrics = []
566
 
567
- training_iter = iter(torch.utils.data.DataLoader(tokenized_datasets.with_format("torch"), batch_size=1, shuffle=False, num_workers=min(33,dataset.n_shards), collate_fn=lambda x: x))
568
 
569
  max_seq_length = min(data_args.max_seq_length, tokenizer.model_max_length)
570
  eval_samples = advance_iter_and_group_samples(training_iter, data_args.num_eval_samples, max_seq_length)
 
564
  train_metrics = []
565
  eval_metrics = []
566
 
567
+ training_iter = iter(torch.utils.data.DataLoader(tokenized_datasets.with_format("torch"), batch_size=1, shuffle=False, num_workers=min(12,dataset.n_shards), collate_fn=lambda x: x))
568
 
569
  max_seq_length = min(data_args.max_seq_length, tokenizer.model_max_length)
570
  eval_samples = advance_iter_and_group_samples(training_iter, data_args.num_eval_samples, max_seq_length)