ttj's picture
Add files using upload-large-folder tool
155e386 verified
raw
history blame
1.15 kB
{"step": 57500, "acc_step": 0, "data_loader_state": {"it_state": {"start_token": 752, "it_state": {"it_state": {"root_dir": "./data", "sources": {"fineweb_edu_10bt_shuffled": 100.0}, "source_to_state": {"fineweb_edu_10bt_shuffled": {"file_path": "data/fineweb_edu_10bt_shuffled/fineweb_edu_10bt.chunk.00.jsonl", "position": 27639381247, "block_size": 4, "offset": 3, "current_iter": 1}}, "rng_state": {"bit_generator": "PCG64", "state": {"state": 286741500864262708623674448239672053530, "inc": 92941856108932518968286621281627530405}, "has_uint32": 0, "uinteger": 0}}, "add_bos": true, "add_eos": true, "name": "tiktoken", "path": "tokenizers/cl100k_base.tiktoken"}, "output_seq_len": 4096, "n_views": 2}, "seq_idx": 312, "rng_state": {"bit_generator": "PCG64", "state": {"state": 240889267456687567804847541390924151647, "inc": 66050176413739185524746886687120723265}, "has_uint32": 1, "uinteger": 2653545164}, "batch_size": 8, "prefetch_size": 1024}, "scheduler": {"base_lrs": [0.003], "last_epoch": 57500, "verbose": false, "_step_count": 57501, "_get_lr_called_within_step": false, "_last_lr": [1.5270821910763795e-05], "lr_lambdas": [{}]}}