hf-transformers-bot commited on
Commit
7477464
·
verified ·
1 Parent(s): eac7ab2

Upload 2025-06-19/runs/63-15763779985/ci_results_run_trainer_and_fsdp_gpu/trainer_and_fsdp_results.json with huggingface_hub

Browse files
2025-06-19/runs/63-15763779985/ci_results_run_trainer_and_fsdp_gpu/trainer_and_fsdp_results.json ADDED
@@ -0,0 +1,204 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "trainer": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 0,
7
+ "multi": 0
8
+ },
9
+ "TensorFlow": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Flax": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Tokenizers": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "Pipelines": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Trainer": {
30
+ "unclassified": 0,
31
+ "single": 3,
32
+ "multi": 2
33
+ },
34
+ "ONNX": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Auto": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ },
44
+ "Quantization": {
45
+ "unclassified": 0,
46
+ "single": 0,
47
+ "multi": 0
48
+ },
49
+ "Unclassified": {
50
+ "unclassified": 0,
51
+ "single": 1,
52
+ "multi": 0
53
+ }
54
+ },
55
+ "success": 399,
56
+ "time_spent": "0:05:29, 0:02:56, ",
57
+ "failures": {
58
+ "multi": [
59
+ {
60
+ "line": "tests/trainer/test_trainer_fsdp.py::TestFSDPTrainerFP8::test_trainer",
61
+ "trace": "(line 2479) RuntimeError: 'accelerate launch --use_fsdp --main_process_port 29500 --num_processes 2 --mixed_precision fp8 --fsdp_transformer_layer_cls_to_wrap GPT2Block /__w/transformers/transformers/tests/trainer/test_trainer_fsdp.py --output_dir /tmp/tmpmfgitrn4 --report_to none --fp8_format HYBRID' failed with returncode 1"
62
+ },
63
+ {
64
+ "line": "tests/trainer/test_trainer_seq2seq.py::Seq2seqTrainerTester::test_return_sequences",
65
+ "trace": "(line 4189) RuntimeError: Graph compile failed. synStatus=synStatus 26 [Generic failure]."
66
+ }
67
+ ],
68
+ "single": [
69
+ {
70
+ "line": "[100%]",
71
+ "trace": "(line 2842) Failed: (subprocess) AssertionError: {'loss': 8.9688, 'grad_norm': nan, 'learnin[53 chars]': 1} != {'epoch': 0.0049261083743842365, 'grad_norm[53 chars]': 1}"
72
+ },
73
+ {
74
+ "line": "tests/trainer/test_trainer.py::TrainerIntegrationTest::test_can_resume_training_lm",
75
+ "trace": "(line 1660) FileNotFoundError: Couldn't find any data file at /__w/transformers/transformers/nyu-mll/glue. Couldn't find 'nyu-mll/glue' on the Hugging Face Hub either: LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on."
76
+ },
77
+ {
78
+ "line": "tests/trainer/test_trainer.py::TrainerIntegrationTest::test_auto_batch_size_finder",
79
+ "trace": "(line 4189) RuntimeError: Graph compile failed. synStatus=synStatus 26 [Generic failure]."
80
+ },
81
+ {
82
+ "line": "tests/trainer/test_trainer_seq2seq.py::Seq2seqTrainerTester::test_return_sequences",
83
+ "trace": "Cannot retrieve error message."
84
+ }
85
+ ]
86
+ },
87
+ "job_link": {
88
+ "multi": "https://github.com/huggingface/transformers/actions/runs/15763779985/job/44436044647",
89
+ "single": "https://github.com/huggingface/transformers/actions/runs/15763779985/job/44436044628"
90
+ }
91
+ },
92
+ "fsdp": {
93
+ "failed": {
94
+ "PyTorch": {
95
+ "unclassified": 0,
96
+ "single": 0,
97
+ "multi": 0
98
+ },
99
+ "TensorFlow": {
100
+ "unclassified": 0,
101
+ "single": 0,
102
+ "multi": 0
103
+ },
104
+ "Flax": {
105
+ "unclassified": 0,
106
+ "single": 0,
107
+ "multi": 0
108
+ },
109
+ "Tokenizers": {
110
+ "unclassified": 0,
111
+ "single": 0,
112
+ "multi": 0
113
+ },
114
+ "Pipelines": {
115
+ "unclassified": 0,
116
+ "single": 0,
117
+ "multi": 0
118
+ },
119
+ "Trainer": {
120
+ "unclassified": 0,
121
+ "single": 0,
122
+ "multi": 0
123
+ },
124
+ "ONNX": {
125
+ "unclassified": 0,
126
+ "single": 0,
127
+ "multi": 0
128
+ },
129
+ "Auto": {
130
+ "unclassified": 0,
131
+ "single": 0,
132
+ "multi": 0
133
+ },
134
+ "Quantization": {
135
+ "unclassified": 0,
136
+ "single": 0,
137
+ "multi": 0
138
+ },
139
+ "Unclassified": {
140
+ "unclassified": 0,
141
+ "single": 0,
142
+ "multi": 12
143
+ }
144
+ },
145
+ "success": 25,
146
+ "time_spent": "0:12:24, .25, ",
147
+ "failures": {
148
+ "multi": [
149
+ {
150
+ "line": "tests/fsdp/test_fsdp.py::TrainerIntegrationFSDP::test_basic_run_full_shard_bf16",
151
+ "trace": "(line 2479) RuntimeError: 'torchrun --nnodes 1 --nproc-per-node 2 --master-port 10999 /__w/transformers/transformers/examples/pytorch/text-classification/run_glue.py --model_name_or_path google-bert/bert-base-cased --task_name mrpc --output_dir /tmp/tmpobqen1r4 --overwrite_output_dir --do_train --max_seq_length 128 --per_device_train_batch_size 16 --learning_rate 5e-5 --num_train_epochs 1 --lr_scheduler_type cosine --logging_steps 50 --save_strategy epoch --do_eval --eval_strategy epoch --report_to none --bf16 --fsdp full_shard auto_wrap --fsdp_transformer_layer_cls_to_wrap BertLayer' failed with returncode 1"
152
+ },
153
+ {
154
+ "line": "tests/fsdp/test_fsdp.py::TrainerIntegrationFSDP::test_basic_run_full_shard_fp16",
155
+ "trace": "(line 2479) RuntimeError: 'torchrun --nnodes 1 --nproc-per-node 2 --master-port 10999 /__w/transformers/transformers/examples/pytorch/text-classification/run_glue.py --model_name_or_path google-bert/bert-base-cased --task_name mrpc --output_dir /tmp/tmpxg9pcy85 --overwrite_output_dir --do_train --max_seq_length 128 --per_device_train_batch_size 16 --learning_rate 5e-5 --num_train_epochs 1 --lr_scheduler_type cosine --logging_steps 50 --save_strategy epoch --do_eval --eval_strategy epoch --report_to none --fp16 --fsdp full_shard auto_wrap --fsdp_transformer_layer_cls_to_wrap BertLayer' failed with returncode 1"
156
+ },
157
+ {
158
+ "line": "tests/fsdp/test_fsdp.py::TrainerIntegrationFSDP::test_basic_run_shard_grad_op_bf16",
159
+ "trace": "(line 2479) RuntimeError: 'torchrun --nnodes 1 --nproc-per-node 2 --master-port 10999 /__w/transformers/transformers/examples/pytorch/text-classification/run_glue.py --model_name_or_path google-bert/bert-base-cased --task_name mrpc --output_dir /tmp/tmpnv7bfxn2 --overwrite_output_dir --do_train --max_seq_length 128 --per_device_train_batch_size 16 --learning_rate 5e-5 --num_train_epochs 1 --lr_scheduler_type cosine --logging_steps 50 --save_strategy epoch --do_eval --eval_strategy epoch --report_to none --bf16 --fsdp shard_grad_op auto_wrap --fsdp_transformer_layer_cls_to_wrap BertLayer' failed with returncode 1"
160
+ },
161
+ {
162
+ "line": "tests/fsdp/test_fsdp.py::TrainerIntegrationFSDP::test_basic_run_shard_grad_op_fp16",
163
+ "trace": "(line 2479) RuntimeError: 'torchrun --nnodes 1 --nproc-per-node 2 --master-port 10999 /__w/transformers/transformers/examples/pytorch/text-classification/run_glue.py --model_name_or_path google-bert/bert-base-cased --task_name mrpc --output_dir /tmp/tmpnpfg7bn8 --overwrite_output_dir --do_train --max_seq_length 128 --per_device_train_batch_size 16 --learning_rate 5e-5 --num_train_epochs 1 --lr_scheduler_type cosine --logging_steps 50 --save_strategy epoch --do_eval --eval_strategy epoch --report_to none --fp16 --fsdp shard_grad_op auto_wrap --fsdp_transformer_layer_cls_to_wrap BertLayer' failed with returncode 1"
164
+ },
165
+ {
166
+ "line": "tests/fsdp/test_fsdp.py::TrainerIntegrationFSDP::test_basic_run_with_cpu_offload_0_bf16",
167
+ "trace": "(line 2479) RuntimeError: 'torchrun --nnodes 1 --nproc-per-node 2 --master-port 10999 /__w/transformers/transformers/examples/pytorch/text-classification/run_glue.py --model_name_or_path google-bert/bert-base-cased --task_name mrpc --output_dir /tmp/tmpg45ldyef --overwrite_output_dir --do_train --max_seq_length 128 --per_device_train_batch_size 16 --learning_rate 5e-5 --num_train_epochs 1 --lr_scheduler_type cosine --logging_steps 50 --save_strategy epoch --do_eval --eval_strategy epoch --report_to none --bf16 --max_steps 10 --fsdp full_shard auto_wrap offload --fsdp_transformer_layer_cls_to_wrap BertLayer' failed with returncode 1"
168
+ },
169
+ {
170
+ "line": "tests/fsdp/test_fsdp.py::TrainerIntegrationFSDP::test_basic_run_with_cpu_offload_1_fp16",
171
+ "trace": "(line 2479) RuntimeError: 'torchrun --nnodes 1 --nproc-per-node 2 --master-port 10999 /__w/transformers/transformers/examples/pytorch/text-classification/run_glue.py --model_name_or_path google-bert/bert-base-cased --task_name mrpc --output_dir /tmp/tmpg7ce67a6 --overwrite_output_dir --do_train --max_seq_length 128 --per_device_train_batch_size 16 --learning_rate 5e-5 --num_train_epochs 1 --lr_scheduler_type cosine --logging_steps 50 --save_strategy epoch --do_eval --eval_strategy epoch --report_to none --fp16 --max_steps 10 --fsdp full_shard auto_wrap offload --fsdp_transformer_layer_cls_to_wrap BertLayer' failed with returncode 1"
172
+ },
173
+ {
174
+ "line": "tests/fsdp/test_fsdp.py::TrainerIntegrationFSDP::test_basic_run_with_gradient_accumulation_full_shard_bf16",
175
+ "trace": "(line 2479) RuntimeError: 'torchrun --nnodes 1 --nproc-per-node 2 --master-port 10999 /__w/transformers/transformers/examples/pytorch/text-classification/run_glue.py --model_name_or_path google-bert/bert-base-cased --task_name mrpc --output_dir /tmp/tmps7mz4rns --overwrite_output_dir --do_train --max_seq_length 128 --per_device_train_batch_size 16 --learning_rate 5e-5 --num_train_epochs 1 --lr_scheduler_type cosine --logging_steps 50 --save_strategy epoch --do_eval --eval_strategy epoch --report_to none --bf16 --gradient_accumulation_steps 2 --fsdp full_shard auto_wrap --fsdp_transformer_layer_cls_to_wrap BertLayer' failed with returncode 1"
176
+ },
177
+ {
178
+ "line": "tests/fsdp/test_fsdp.py::TrainerIntegrationFSDP::test_basic_run_with_gradient_accumulation_full_shard_fp16",
179
+ "trace": "(line 2479) RuntimeError: 'torchrun --nnodes 1 --nproc-per-node 2 --master-port 10999 /__w/transformers/transformers/examples/pytorch/text-classification/run_glue.py --model_name_or_path google-bert/bert-base-cased --task_name mrpc --output_dir /tmp/tmpjo22y3ve --overwrite_output_dir --do_train --max_seq_length 128 --per_device_train_batch_size 16 --learning_rate 5e-5 --num_train_epochs 1 --lr_scheduler_type cosine --logging_steps 50 --save_strategy epoch --do_eval --eval_strategy epoch --report_to none --fp16 --gradient_accumulation_steps 2 --fsdp full_shard auto_wrap --fsdp_transformer_layer_cls_to_wrap BertLayer' failed with returncode 1"
180
+ },
181
+ {
182
+ "line": "tests/fsdp/test_fsdp.py::TrainerIntegrationFSDP::test_basic_run_with_gradient_accumulation_shard_grad_op_bf16",
183
+ "trace": "(line 2479) RuntimeError: 'torchrun --nnodes 1 --nproc-per-node 2 --master-port 10999 /__w/transformers/transformers/examples/pytorch/text-classification/run_glue.py --model_name_or_path google-bert/bert-base-cased --task_name mrpc --output_dir /tmp/tmpi39sr7b5 --overwrite_output_dir --do_train --max_seq_length 128 --per_device_train_batch_size 16 --learning_rate 5e-5 --num_train_epochs 1 --lr_scheduler_type cosine --logging_steps 50 --save_strategy epoch --do_eval --eval_strategy epoch --report_to none --bf16 --gradient_accumulation_steps 2 --fsdp shard_grad_op auto_wrap --fsdp_transformer_layer_cls_to_wrap BertLayer' failed with returncode 1"
184
+ },
185
+ {
186
+ "line": "tests/fsdp/test_fsdp.py::TrainerIntegrationFSDP::test_basic_run_with_gradient_accumulation_shard_grad_op_fp16",
187
+ "trace": "(line 2479) RuntimeError: 'torchrun --nnodes 1 --nproc-per-node 2 --master-port 10999 /__w/transformers/transformers/examples/pytorch/text-classification/run_glue.py --model_name_or_path google-bert/bert-base-cased --task_name mrpc --output_dir /tmp/tmpvu4rt_0f --overwrite_output_dir --do_train --max_seq_length 128 --per_device_train_batch_size 16 --learning_rate 5e-5 --num_train_epochs 1 --lr_scheduler_type cosine --logging_steps 50 --save_strategy epoch --do_eval --eval_strategy epoch --report_to none --fp16 --gradient_accumulation_steps 2 --fsdp shard_grad_op auto_wrap --fsdp_transformer_layer_cls_to_wrap BertLayer' failed with returncode 1"
188
+ },
189
+ {
190
+ "line": "tests/fsdp/test_fsdp.py::TrainerIntegrationFSDP::test_training_and_can_resume_normally_FULL_STATE_DICT",
191
+ "trace": "(line 2479) RuntimeError: 'torchrun --nnodes 1 --nproc-per-node 2 --master-port 10999 /__w/transformers/transformers/examples/pytorch/text-classification/run_glue.py --model_name_or_path google-bert/bert-base-cased --task_name mrpc --output_dir ./xxx --overwrite_output_dir --do_train --max_seq_length 128 --per_device_train_batch_size 16 --learning_rate 5e-5 --num_train_epochs 2 --lr_scheduler_type cosine --logging_steps 25 --save_strategy epoch --do_eval --eval_strategy epoch --report_to none --fsdp full_shard auto_wrap --fsdp_transformer_layer_cls_to_wrap BertLayer' failed with returncode 1"
192
+ },
193
+ {
194
+ "line": "tests/fsdp/test_fsdp.py::TrainerIntegrationFSDP::test_training_and_can_resume_normally_SHARDED_STATE_DICT",
195
+ "trace": "(line 2479) RuntimeError: 'accelerate launch --num_processes 2 --main_process_port 10999 --use_fsdp --fsdp_auto_wrap_policy TRANSFORMER_BASED_WRAP --fsdp_state_dict_type SHARDED_STATE_DICT --fsdp_transformer_layer_cls_to_wrap BertLayer --fsdp_sharding_strategy 1 /__w/transformers/transformers/examples/pytorch/text-classification/run_glue.py --model_name_or_path google-bert/bert-base-cased --task_name mrpc --output_dir ./xxx --overwrite_output_dir --do_train --max_seq_length 128 --per_device_train_batch_size 16 --learning_rate 5e-5 --num_train_epochs 2 --lr_scheduler_type cosine --logging_steps 25 --save_strategy epoch --do_eval --eval_strategy epoch --report_to none' failed with returncode 1"
196
+ }
197
+ ]
198
+ },
199
+ "job_link": {
200
+ "multi": "https://github.com/huggingface/transformers/actions/runs/15763779985/job/44436044632",
201
+ "single": "https://github.com/huggingface/transformers/actions/runs/15763779985/job/44436044623"
202
+ }
203
+ }
204
+ }