# Running on cnode7-008 # Started at Fri Nov 29 11:08:40 CST 2024 # SLURMD_NODENAME=cnode7-008 # SLURM_CLUSTER_NAME=slurm # SLURM_CONF=/cm/shared/apps/slurm/var/etc/slurm/slurm.conf # SLURM_CPUS_ON_NODE=224 # SLURM_CPUS_PER_TASK=8 # SLURM_EXPORT_ENV=PATH # SLURM_GET_USER_ENV=1 # SLURM_GPUS_ON_NODE=8 # SLURM_GPUS_PER_NODE=8 # SLURM_GTIDS=0 # SLURM_JOBID=4964 # SLURM_JOB_CPUS_PER_NODE='224(x2)' # SLURM_JOB_END_TIME=1764385707 # SLURM_JOB_GID=1026 # SLURM_JOB_GPUS=0,1,2,3,4,5,6,7 # SLURM_JOB_ID=4964 # SLURM_JOB_NAME=exp_owsm/s2t_train_05b_ds_raw_bpe50000/train.log # SLURM_JOB_NODELIST='cnode7-[008-009]' # SLURM_JOB_NUM_NODES=2 # SLURM_JOB_PARTITION=p2 # SLURM_JOB_QOS=normal # SLURM_JOB_START_TIME=1732849707 # SLURM_JOB_UID=1026 # SLURM_JOB_USER=williamchen # SLURM_LOCALID=0 # SLURM_MEM_PER_NODE=2048000 # SLURM_NNODES=2 # SLURM_NODEID=0 # SLURM_NODELIST='cnode7-[008-009]' # SLURM_NODE_ALIASES='(null)' # SLURM_OPEN_MODE=a # SLURM_PRIO_PROCESS=0 # SLURM_PROCID=0 # SLURM_SUBMIT_DIR=/mnt/home/williamchen/espnet/egs2/owsm_v3.1/s2t1 # SLURM_SUBMIT_HOST=154-T2-P1-NVR # SLURM_TASKS_PER_NODE='28(x2)' # SLURM_TASK_PID=3115096 # SLURM_TOPOLOGY_ADDR=cnode7-008 # SLURM_TOPOLOGY_ADDR_PATTERN=node # SLURM_WORKING_CLUSTER=slurm:154-T2-P1-NVR:6817:9984:109 # srun --export=ALL python3 -m espnet2.bin.s2t_train --use_preprocessor true --bpemodel data/token_list/bpe_unigram50000/bpe.model --token_type bpe --token_list data/token_list/bpe_unigram50000/tokens.txt --non_linguistic_symbols none --cleaner none --g2p none --valid_data_path_and_name_and_type dump/raw/dev_v3/wav.scp,speech,kaldi_ark --valid_shape_file exp_owsm/s2t_stats_raw_bpe50000/valid/speech_shape --resume true --fold_length 80000 --output_dir exp_owsm/s2t_train_05b_ds_raw_bpe50000 --config conf/train_05b_ds.yaml --frontend_conf fs=16k --normalize=global_mvn --normalize_conf stats_file=exp_owsm/s2t_stats_raw_bpe50000/train/feats_stats.npz --train_data_path_and_name_and_type exp_owsm/s2t_stats_raw_bpe50000/splits8/wav.scp,speech,kaldi_ark --train_shape_file exp_owsm/s2t_stats_raw_bpe50000/splits8/speech_shape --fold_length 150 --train_data_path_and_name_and_type exp_owsm/s2t_stats_raw_bpe50000/splits8/text.prev,text_prev,text --fold_length 150 --train_data_path_and_name_and_type exp_owsm/s2t_stats_raw_bpe50000/splits8/text.ctc,text_ctc,text --fold_length 150 --train_data_path_and_name_and_type exp_owsm/s2t_stats_raw_bpe50000/splits8/text,text,text --multiple_iterator true --valid_data_path_and_name_and_type dump/raw/dev_v3/text.prev,text_prev,text --valid_shape_file exp_owsm/s2t_stats_raw_bpe50000/valid/text_prev_shape --valid_data_path_and_name_and_type dump/raw/dev_v3/text.ctc,text_ctc,text --valid_shape_file exp_owsm/s2t_stats_raw_bpe50000/valid/text_ctc_shape --valid_data_path_and_name_and_type dump/raw/dev_v3/text,text,text --valid_shape_file exp_owsm/s2t_stats_raw_bpe50000/valid/text_shape --ngpu 8 --multiprocessing_distributed true --dist_launcher slurm --dist_init_method file:///mnt/home/williamchen/espnet/egs2/owsm_v3.1/s2t1/exp_owsm/s2t_train_05b_ds_raw_bpe50000/.dist_init_e725dc5e-341f-47ac-9ef3-02e7977753e5 [2024-11-29 11:08:58,541] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:08:45,920] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/bin/python3 /mnt/home/williamchen/espnet/espnet2/bin/s2t_train.py --use_preprocessor true --bpemodel data/token_list/bpe_unigram50000/bpe.model --token_type bpe --token_list data/token_list/bpe_unigram50000/tokens.txt --non_linguistic_symbols none --cleaner none --g2p none --valid_data_path_and_name_and_type dump/raw/dev_v3/wav.scp,speech,kaldi_ark --valid_shape_file exp_owsm/s2t_stats_raw_bpe50000/valid/speech_shape --resume true --fold_length 80000 --output_dir exp_owsm/s2t_train_05b_ds_raw_bpe50000 --config conf/train_05b_ds.yaml --frontend_conf fs=16k --normalize=global_mvn --normalize_conf stats_file=exp_owsm/s2t_stats_raw_bpe50000/train/feats_stats.npz --train_data_path_and_name_and_type exp_owsm/s2t_stats_raw_bpe50000/splits8/wav.scp,speech,kaldi_ark --train_shape_file exp_owsm/s2t_stats_raw_bpe50000/splits8/speech_shape --fold_length 150 --train_data_path_and_name_and_type exp_owsm/s2t_stats_raw_bpe50000/splits8/text.prev,text_prev,text --fold_length 150 --train_data_path_and_name_and_type exp_owsm/s2t_stats_raw_bpe50000/splits8/text.ctc,text_ctc,text --fold_length 150 --train_data_path_and_name_and_type exp_owsm/s2t_stats_raw_bpe50000/splits8/text,text,text --multiple_iterator true --valid_data_path_and_name_and_type dump/raw/dev_v3/text.prev,text_prev,text --valid_shape_file exp_owsm/s2t_stats_raw_bpe50000/valid/text_prev_shape --valid_data_path_and_name_and_type dump/raw/dev_v3/text.ctc,text_ctc,text --valid_shape_file exp_owsm/s2t_stats_raw_bpe50000/valid/text_ctc_shape --valid_data_path_and_name_and_type dump/raw/dev_v3/text,text,text --valid_shape_file exp_owsm/s2t_stats_raw_bpe50000/valid/text_shape --ngpu 8 --multiprocessing_distributed true --dist_launcher slurm --dist_init_method file:///mnt/home/williamchen/espnet/egs2/owsm_v3.1/s2t1/exp_owsm/s2t_train_05b_ds_raw_bpe50000/.dist_init_e725dc5e-341f-47ac-9ef3-02e7977753e5 /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/bin/python3 /mnt/home/williamchen/espnet/espnet2/bin/s2t_train.py --use_preprocessor true --bpemodel data/token_list/bpe_unigram50000/bpe.model --token_type bpe --token_list data/token_list/bpe_unigram50000/tokens.txt --non_linguistic_symbols none --cleaner none --g2p none --valid_data_path_and_name_and_type dump/raw/dev_v3/wav.scp,speech,kaldi_ark --valid_shape_file exp_owsm/s2t_stats_raw_bpe50000/valid/speech_shape --resume true --fold_length 80000 --output_dir exp_owsm/s2t_train_05b_ds_raw_bpe50000 --config conf/train_05b_ds.yaml --frontend_conf fs=16k --normalize=global_mvn --normalize_conf stats_file=exp_owsm/s2t_stats_raw_bpe50000/train/feats_stats.npz --train_data_path_and_name_and_type exp_owsm/s2t_stats_raw_bpe50000/splits8/wav.scp,speech,kaldi_ark --train_shape_file exp_owsm/s2t_stats_raw_bpe50000/splits8/speech_shape --fold_length 150 --train_data_path_and_name_and_type exp_owsm/s2t_stats_raw_bpe50000/splits8/text.prev,text_prev,text --fold_length 150 --train_data_path_and_name_and_type exp_owsm/s2t_stats_raw_bpe50000/splits8/text.ctc,text_ctc,text --fold_length 150 --train_data_path_and_name_and_type exp_owsm/s2t_stats_raw_bpe50000/splits8/text,text,text --multiple_iterator true --valid_data_path_and_name_and_type dump/raw/dev_v3/text.prev,text_prev,text --valid_shape_file exp_owsm/s2t_stats_raw_bpe50000/valid/text_prev_shape --valid_data_path_and_name_and_type dump/raw/dev_v3/text.ctc,text_ctc,text --valid_shape_file exp_owsm/s2t_stats_raw_bpe50000/valid/text_ctc_shape --valid_data_path_and_name_and_type dump/raw/dev_v3/text,text,text --valid_shape_file exp_owsm/s2t_stats_raw_bpe50000/valid/text_shape --ngpu 8 --multiprocessing_distributed true --dist_launcher slurm --dist_init_method file:///mnt/home/williamchen/espnet/egs2/owsm_v3.1/s2t1/exp_owsm/s2t_train_05b_ds_raw_bpe50000/.dist_init_e725dc5e-341f-47ac-9ef3-02e7977753e5 [2024-11-29 11:09:25,554] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:25,707] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:26,295] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:26,695] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:26,717] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:26,849] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:26,854] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:26,891] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:14,662] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:14,733] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:15,016] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:15,286] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:15,306] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:15,391] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:15,414] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:09:15,436] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [W1129 11:09:34.193793407 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:34.299097392 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:35.856913182 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:35.304748753 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:36.358830097 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:36.568232324 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:36.576747309 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [cnode7-008:0/16] 2024-11-29 11:09:36,254 (s2t:462) INFO: Vocabulary size: 50002 [W1129 11:09:36.598399867 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:24.396105211 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:24.413895338 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:24.414229307 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:24.922594070 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:24.928493435 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:24.985020333 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:25.263124355 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [W1129 11:09:25.303328532 Utils.hpp:164] Warning: Environment variable NCCL_BLOCKING_WAIT is deprecated; use TORCH_NCCL_BLOCKING_WAIT instead (function operator()) [cnode7-008:0/16] 2024-11-29 11:09:39,954 (abs_task:1383) INFO: pytorch.version=2.4.0+cu121, cuda.available=True, cudnn.version=90100, cudnn.benchmark=False, cudnn.deterministic=True [cnode7-008:0/16] 2024-11-29 11:09:39,960 (abs_task:1384) INFO: Model structure: ESPnetS2TModel( (frontend): DefaultFrontend( (stft): Stft(n_fft=512, win_length=400, hop_length=160, center=True, normalized=False, onesided=True) (frontend): Frontend() (logmel): LogMel(sr=16000, n_fft=512, n_mels=80, fmin=0, fmax=8000.0, htk=False) ) (specaug): SpecAug( (freq_mask): MaskAlongAxis(mask_width_range=[0, 27], num_mask=2, axis=freq) (time_mask): MaskAlongAxisVariableMaxWidth(mask_width_ratio_range=[0.0, 0.05], num_mask=10, axis=time) ) (normalize): GlobalMVN(stats_file=exp_owsm/s2t_stats_raw_bpe50000/train/feats_stats.npz, norm_means=True, norm_vars=True) (encoder): TransformerEncoder( (embed): Conv2dSubsampling( (conv): Sequential( (0): Conv2d(1, 1024, kernel_size=(3, 3), stride=(2, 2)) (1): ReLU() (2): Conv2d(1024, 1024, kernel_size=(3, 3), stride=(2, 2)) (3): ReLU() ) (out): Sequential( (0): Linear(in_features=19456, out_features=1024, bias=True) (1): PositionalEncoding( (dropout): Dropout(p=0.1, inplace=False) ) ) ) (encoders): MultiSequential( (0): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (1): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (2): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (3): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (4): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (5): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (6): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (7): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (8): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (9): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (10): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (11): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (12): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (13): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (14): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (15): EncoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) ) (after_norm): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) ) (decoder): TransformerDecoder( (embed): Sequential( (0): Embedding(50002, 1024) (1): PositionalEncoding( (dropout): Dropout(p=0.1, inplace=False) ) ) (after_norm): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (output_layer): Linear(in_features=1024, out_features=50002, bias=True) (decoders): MultiSequential( (0): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (1): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (2): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (3): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (4): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (5): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (6): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (7): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (8): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (9): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (10): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (11): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (12): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (13): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (14): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) (15): DecoderLayer( (self_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (src_attn): MultiHeadedAttention( (linear_q): Linear(in_features=1024, out_features=1024, bias=True) (linear_k): Linear(in_features=1024, out_features=1024, bias=True) (linear_v): Linear(in_features=1024, out_features=1024, bias=True) (linear_out): Linear(in_features=1024, out_features=1024, bias=True) (dropout): Identity() (q_norm): Identity() (k_norm): Identity() ) (feed_forward): PositionwiseFeedForward( (w_1): Linear(in_features=1024, out_features=4096, bias=True) (w_2): Linear(in_features=4096, out_features=1024, bias=True) (dropout): Dropout(p=0.1, inplace=False) (activation): ReLU() ) (norm1): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm2): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (norm3): LayerNorm((1024,), eps=1e-12, elementwise_affine=True) (dropout): Dropout(p=0.1, inplace=False) ) ) ) (criterion_att): LabelSmoothingLoss( (criterion): KLDivLoss() ) (ctc): CTC( (ctc_lo): Linear(in_features=1024, out_features=50002, bias=True) (ctc_loss): CTCLoss() ) ) Model summary: Class Name: ESPnetS2TModel Total Number of model parameters: 653.37 M Number of trainable parameters: 653.37 M (100.0%) Size: 2.61 GB Type: torch.float32 [cnode7-008:0/16] 2024-11-29 11:09:39,960 (abs_task:1387) INFO: Optimizer: Adadelta ( Parameter Group 0 capturable: False differentiable: False eps: 1e-06 foreach: None lr: 1.0 maximize: False rho: 0.9 weight_decay: 0 ) [cnode7-008:0/16] 2024-11-29 11:09:39,960 (abs_task:1388) INFO: Scheduler: None [cnode7-008:0/16] 2024-11-29 11:09:39,974 (abs_task:1397) INFO: Saving the configuration in exp_owsm/s2t_train_05b_ds_raw_bpe50000/config.yaml /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( [2024-11-29 11:09:44,564] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:44,572] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( [2024-11-29 11:09:44,619] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:44,621] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 [2024-11-29 11:09:44,621] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:44,622] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 [cnode7-008:0/16] 2024-11-29 11:09:44,658 (s2t:444) INFO: Optional Data Names: ('text_prev', 'text_ctc', 'text_spk2', 'text_spk3', 'text_spk4') /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( [2024-11-29 11:09:45,331] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:45,331] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:45,332] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 [2024-11-29 11:09:45,332] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( [2024-11-29 11:09:45,336] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:45,336] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( [2024-11-29 11:09:45,460] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:45,461] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 [cnode7-008:0/16] 2024-11-29 11:09:45,573 (abs_task:1807) INFO: [valid] dataset: ESPnetDataset( speech: {"path": "dump/raw/dev_v3/wav.scp", "type": "kaldi_ark"} text_prev: {"path": "dump/raw/dev_v3/text.prev", "type": "text"} text_ctc: {"path": "dump/raw/dev_v3/text.ctc", "type": "text"} text: {"path": "dump/raw/dev_v3/text", "type": "text"} preprocess: ) [cnode7-008:0/16] 2024-11-29 11:09:45,573 (abs_task:1808) INFO: [valid] Batch sampler: UnsortedBatchSampler(N-batch=74743, batch_size=16, key_file=exp_owsm/s2t_stats_raw_bpe50000/valid/speech_shape, [cnode7-008:0/16] 2024-11-29 11:09:45,579 (abs_task:1809) INFO: [valid] mini-batch sizes summary: N-batch=74743, mean=16.0, min=16, max=17 /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( [cnode7-008:0/16] 2024-11-29 11:09:45,827 (distributed_utils:129) WARNING: ================================================================= Found OMP_NUM_THREADS=1 in environment variables. With some advanced features, DeepSpeed may have heavy cpu workload so that OMP_NUM_THREADS=1 is not sufficient. Try to increase it in your path.sh ================================================================= [2024-11-29 11:09:45,827] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:45,828] [INFO] [logging.py:129:log_dist] [Rank 0] DeepSpeed info: version=0.15.3, git-hash=unknown, git-branch=unknown [2024-11-29 11:09:45,828] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 cnode7-008:3116459:3116459 [0] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116459:3116459 [0] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.170<0> cnode7-008:3116459:3116459 [0] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-008:3116459:3116459 [0] NCCL INFO cudaDriverVersion 12020 NCCL version 2.20.5+cuda12.4 cnode7-008:3116465:3116465 [6] NCCL INFO cudaDriverVersion 12020 cnode7-008:3116465:3116465 [6] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116465:3116465 [6] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.170<0> cnode7-008:3116465:3116465 [6] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-008:3116463:3116463 [4] NCCL INFO cudaDriverVersion 12020 cnode7-008:3116463:3116463 [4] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116463:3116463 [4] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.170<0> cnode7-008:3116463:3116463 [4] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-008:3116460:3116460 [1] NCCL INFO cudaDriverVersion 12020 cnode7-008:3116460:3116460 [1] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116460:3116460 [1] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.170<0> cnode7-008:3116460:3116460 [1] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-008:3116461:3116461 [2] NCCL INFO cudaDriverVersion 12020 cnode7-008:3116461:3116461 [2] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116461:3116461 [2] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.170<0> cnode7-008:3116461:3116461 [2] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-008:3116466:3116466 [7] NCCL INFO cudaDriverVersion 12020 cnode7-008:3116466:3116466 [7] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116466:3116466 [7] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.170<0> cnode7-008:3116466:3116466 [7] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-008:3116462:3116462 [3] NCCL INFO cudaDriverVersion 12020 cnode7-008:3116462:3116462 [3] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116462:3116462 [3] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.170<0> cnode7-008:3116462:3116462 [3] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-008:3116464:3116464 [5] NCCL INFO cudaDriverVersion 12020 cnode7-008:3116464:3116464 [5] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116464:3116464 [5] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.170<0> cnode7-008:3116464:3116464 [5] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-008:3116464:3117204 [5] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116459:3117197 [0] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116465:3117199 [6] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116463:3117200 [4] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116463:3117200 [4] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.170<0> cnode7-008:3116463:3117200 [4] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116463:3117200 [4] NCCL INFO Using network IB cnode7-008:3116460:3117203 [1] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116461:3117202 [2] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116466:3117198 [7] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116462:3117205 [3] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-008:3116462:3117205 [3] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.170<0> cnode7-008:3116462:3117205 [3] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116462:3117205 [3] NCCL INFO Using network IB cnode7-008:3116464:3117204 [5] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.170<0> cnode7-008:3116464:3117204 [5] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116464:3117204 [5] NCCL INFO Using network IB cnode7-008:3116459:3117197 [0] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.170<0> cnode7-008:3116459:3117197 [0] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116459:3117197 [0] NCCL INFO Using network IB cnode7-008:3116465:3117199 [6] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.170<0> cnode7-008:3116465:3117199 [6] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116465:3117199 [6] NCCL INFO Using network IB cnode7-008:3116460:3117203 [1] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.170<0> cnode7-008:3116460:3117203 [1] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116460:3117203 [1] NCCL INFO Using network IB cnode7-008:3116461:3117202 [2] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.170<0> cnode7-008:3116461:3117202 [2] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116461:3117202 [2] NCCL INFO Using network IB cnode7-008:3116466:3117198 [7] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.170<0> cnode7-008:3116466:3117198 [7] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116466:3117198 [7] NCCL INFO Using network IB /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( [2024-11-29 11:09:33,982] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:33,984] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:33,984] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:33,989] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 [2024-11-29 11:09:33,989] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 [2024-11-29 11:09:33,989] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( [2024-11-29 11:09:34,010] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:34,012] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( [2024-11-29 11:09:34,072] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:34,073] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 cnode7-009:2307519:2307519 [0] NCCL INFO cudaDriverVersion 12020 cnode7-009:2307519:2307519 [0] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307519:2307519 [0] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.171<0> cnode7-009:2307519:2307519 [0] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-009:2307525:2307525 [6] NCCL INFO cudaDriverVersion 12020 cnode7-009:2307525:2307525 [6] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307525:2307525 [6] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.171<0> cnode7-009:2307525:2307525 [6] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-009:2307523:2307523 [4] NCCL INFO cudaDriverVersion 12020 cnode7-009:2307523:2307523 [4] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307523:2307523 [4] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.171<0> cnode7-009:2307523:2307523 [4] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-009:2307526:2307526 [7] NCCL INFO cudaDriverVersion 12020 cnode7-009:2307526:2307526 [7] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307526:2307526 [7] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.171<0> cnode7-009:2307526:2307526 [7] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-009:2307521:2307521 [2] NCCL INFO cudaDriverVersion 12020 cnode7-009:2307521:2307521 [2] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307521:2307521 [2] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.171<0> cnode7-009:2307521:2307521 [2] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( [2024-11-29 11:09:34,276] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:34,278] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 cnode7-009:2307525:2308209 [6] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307523:2308206 [4] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307526:2308203 [7] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307521:2308204 [2] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307522:2307522 [3] NCCL INFO cudaDriverVersion 12020 cnode7-009:2307522:2307522 [3] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307522:2307522 [3] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.171<0> cnode7-009:2307522:2307522 [3] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-009:2307519:2308205 [0] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( [2024-11-29 11:09:34,461] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:34,462] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 cnode7-009:2307525:2308209 [6] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.171<0> cnode7-009:2307525:2308209 [6] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307525:2308209 [6] NCCL INFO Using network IB cnode7-009:2307523:2308206 [4] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.171<0> cnode7-009:2307523:2308206 [4] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307523:2308206 [4] NCCL INFO Using network IB cnode7-009:2307526:2308203 [7] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.171<0> cnode7-009:2307526:2308203 [7] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307526:2308203 [7] NCCL INFO Using network IB cnode7-009:2307521:2308204 [2] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.171<0> cnode7-009:2307521:2308204 [2] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307521:2308204 [2] NCCL INFO Using network IB cnode7-009:2307519:2308205 [0] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.171<0> cnode7-009:2307519:2308205 [0] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307519:2308205 [0] NCCL INFO Using network IB /mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py:240: InstrumentationWarning: @typechecked only supports instrumenting functions wrapped with @classmethod, @staticmethod or @property -- not typechecking espnet2.train.deepspeed_trainer.DeepSpeedTrainer.valid_one_epoch def valid_one_epoch( [2024-11-29 11:09:34,654] [INFO] [comm.py:652:init_distributed] cdb=None [2024-11-29 11:09:34,655] [INFO] [config.py:733:__init__] Config mesh_device None world_size = 16 cnode7-009:2307522:2308212 [3] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307522:2308212 [3] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.171<0> cnode7-009:2307522:2308212 [3] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307522:2308212 [3] NCCL INFO Using network IB cnode7-009:2307520:2307520 [1] NCCL INFO cudaDriverVersion 12020 cnode7-009:2307520:2307520 [1] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307520:2307520 [1] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.171<0> cnode7-009:2307520:2307520 [1] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-009:2307524:2307524 [5] NCCL INFO cudaDriverVersion 12020 cnode7-009:2307524:2307524 [5] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307524:2307524 [5] NCCL INFO Bootstrap : Using enp170s0np0:10.225.1.171<0> cnode7-009:2307524:2307524 [5] NCCL INFO NET/Plugin : dlerror=libnccl-net.so: cannot open shared object file: No such file or directory No plugin found (libnccl-net.so), using internal implementation cnode7-009:2307520:2308272 [1] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307520:2308272 [1] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.171<0> cnode7-009:2307520:2308272 [1] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307520:2308272 [1] NCCL INFO Using network IB cnode7-009:2307524:2308273 [5] NCCL INFO NCCL_SOCKET_IFNAME set by environment to ^lo,docker,virbr,vmnet,vboxnet cnode7-009:2307524:2308273 [5] NCCL INFO NET/IB : Using [0]mlx5_0:1/IB [1]mlx5_3:1/IB [2]mlx5_4:1/IB [3]mlx5_5:1/IB [4]mlx5_6:1/IB [5]mlx5_7:1/RoCE [6]mlx5_8:1/IB [7]mlx5_9:1/IB [8]mlx5_10:1/IB [RO]; OOB enp170s0np0:10.225.1.171<0> cnode7-009:2307524:2308273 [5] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307524:2308273 [5] NCCL INFO Using network IB cnode7-008:3116465:3117199 [6] NCCL INFO comm 0x555560460a80 rank 6 nranks 16 cudaDev 6 nvmlDev 6 busId d1000 commId 0x821cf94b50e14dd2 - Init START cnode7-008:3116463:3117200 [4] NCCL INFO comm 0x55557bc1dc20 rank 4 nranks 16 cudaDev 4 nvmlDev 4 busId 9d000 commId 0x821cf94b50e14dd2 - Init START cnode7-008:3116460:3117203 [1] NCCL INFO comm 0x55558101c4b0 rank 1 nranks 16 cudaDev 1 nvmlDev 1 busId 43000 commId 0x821cf94b50e14dd2 - Init START cnode7-009:2307524:2308273 [5] NCCL INFO comm 0x555560098c20 rank 13 nranks 16 cudaDev 5 nvmlDev 5 busId c3000 commId 0x821cf94b50e14dd2 - Init START cnode7-008:3116461:3117202 [2] NCCL INFO comm 0x5555808662f0 rank 2 nranks 16 cudaDev 2 nvmlDev 2 busId 52000 commId 0x821cf94b50e14dd2 - Init START cnode7-009:2307519:2308205 [0] NCCL INFO comm 0x555564bcae80 rank 8 nranks 16 cudaDev 0 nvmlDev 0 busId 1b000 commId 0x821cf94b50e14dd2 - Init START cnode7-008:3116466:3117198 [7] NCCL INFO comm 0x555560c1ca30 rank 7 nranks 16 cudaDev 7 nvmlDev 7 busId df000 commId 0x821cf94b50e14dd2 - Init START cnode7-008:3116462:3117205 [3] NCCL INFO comm 0x5555810234b0 rank 3 nranks 16 cudaDev 3 nvmlDev 3 busId 61000 commId 0x821cf94b50e14dd2 - Init START cnode7-009:2307525:2308209 [6] NCCL INFO comm 0x555567c1b8e0 rank 14 nranks 16 cudaDev 6 nvmlDev 6 busId d1000 commId 0x821cf94b50e14dd2 - Init START cnode7-009:2307523:2308206 [4] NCCL INFO comm 0x55556005a740 rank 12 nranks 16 cudaDev 4 nvmlDev 4 busId 9d000 commId 0x821cf94b50e14dd2 - Init START cnode7-008:3116464:3117204 [5] NCCL INFO comm 0x555575421c00 rank 5 nranks 16 cudaDev 5 nvmlDev 5 busId c3000 commId 0x821cf94b50e14dd2 - Init START cnode7-008:3116459:3117197 [0] NCCL INFO comm 0x5555647ca110 rank 0 nranks 16 cudaDev 0 nvmlDev 0 busId 1b000 commId 0x821cf94b50e14dd2 - Init START cnode7-009:2307521:2308204 [2] NCCL INFO comm 0x5555810227a0 rank 10 nranks 16 cudaDev 2 nvmlDev 2 busId 52000 commId 0x821cf94b50e14dd2 - Init START cnode7-009:2307526:2308203 [7] NCCL INFO comm 0x555560a16e20 rank 15 nranks 16 cudaDev 7 nvmlDev 7 busId df000 commId 0x821cf94b50e14dd2 - Init START cnode7-009:2307522:2308212 [3] NCCL INFO comm 0x555580c22f60 rank 11 nranks 16 cudaDev 3 nvmlDev 3 busId 61000 commId 0x821cf94b50e14dd2 - Init START cnode7-009:2307520:2308272 [1] NCCL INFO comm 0x555560a202e0 rank 9 nranks 16 cudaDev 1 nvmlDev 1 busId 43000 commId 0x821cf94b50e14dd2 - Init START cnode7-008:3116462:3117205 [3] NCCL INFO Setting affinity for GPU 3 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-008:3116462:3117205 [3] NCCL INFO NVLS multicast support is available on dev 3 cnode7-009:2307524:2308273 [5] NCCL INFO Setting affinity for GPU 5 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-009:2307524:2308273 [5] NCCL INFO NVLS multicast support is available on dev 5 cnode7-008:3116464:3117204 [5] NCCL INFO Setting affinity for GPU 5 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-008:3116464:3117204 [5] NCCL INFO NVLS multicast support is available on dev 5 cnode7-008:3116459:3117197 [0] NCCL INFO Setting affinity for GPU 0 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-008:3116459:3117197 [0] NCCL INFO NVLS multicast support is available on dev 0 cnode7-009:2307526:2308203 [7] NCCL INFO Setting affinity for GPU 7 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-009:2307526:2308203 [7] NCCL INFO NVLS multicast support is available on dev 7 cnode7-008:3116465:3117199 [6] NCCL INFO Setting affinity for GPU 6 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-008:3116465:3117199 [6] NCCL INFO NVLS multicast support is available on dev 6 cnode7-008:3116461:3117202 [2] NCCL INFO Setting affinity for GPU 2 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-008:3116461:3117202 [2] NCCL INFO NVLS multicast support is available on dev 2 cnode7-008:3116466:3117198 [7] NCCL INFO Setting affinity for GPU 7 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-008:3116466:3117198 [7] NCCL INFO NVLS multicast support is available on dev 7 cnode7-008:3116463:3117200 [4] NCCL INFO Setting affinity for GPU 4 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-008:3116463:3117200 [4] NCCL INFO NVLS multicast support is available on dev 4 cnode7-008:3116460:3117203 [1] NCCL INFO Setting affinity for GPU 1 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-008:3116460:3117203 [1] NCCL INFO NVLS multicast support is available on dev 1 cnode7-009:2307521:2308204 [2] NCCL INFO Setting affinity for GPU 2 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-009:2307521:2308204 [2] NCCL INFO NVLS multicast support is available on dev 2 cnode7-009:2307522:2308212 [3] NCCL INFO Setting affinity for GPU 3 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-009:2307522:2308212 [3] NCCL INFO NVLS multicast support is available on dev 3 cnode7-009:2307520:2308272 [1] NCCL INFO Setting affinity for GPU 1 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-009:2307520:2308272 [1] NCCL INFO NVLS multicast support is available on dev 1 cnode7-009:2307519:2308205 [0] NCCL INFO Setting affinity for GPU 0 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-009:2307519:2308205 [0] NCCL INFO NVLS multicast support is available on dev 0 cnode7-009:2307521:2308204 [2] NCCL INFO comm 0x5555810227a0 rank 10 nRanks 16 nNodes 2 localRanks 8 localRank 2 MNNVL 0 cnode7-009:2307521:2308204 [2] NCCL INFO Trees [0] 11/-1/-1->10->9 [1] 11/-1/-1->10->9 [2] 11/-1/-1->10->2 [3] -1/-1/-1->10->9 [4] 11/-1/-1->10->9 [5] 11/-1/-1->10->9 [6] 11/-1/-1->10->9 [7] 11/-1/-1->10->9 [8] 11/-1/-1->10->9 [9] 11/-1/-1->10->9 [10] 11/2/-1->10->-1 [11] -1/-1/-1->10->9 [12] 11/-1/-1->10->9 [13] 11/-1/-1->10->9 [14] 11/-1/-1->10->9 [15] 11/-1/-1->10->9 cnode7-009:2307521:2308204 [2] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116465:3117199 [6] NCCL INFO comm 0x555560460a80 rank 6 nRanks 16 nNodes 2 localRanks 8 localRank 6 MNNVL 0 cnode7-008:3116465:3117199 [6] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116465:3117199 [6] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116465:3117199 [6] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116465:3117199 [6] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116465:3117199 [6] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116465:3117199 [6] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116465:3117199 [6] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116465:3117199 [6] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116465:3117199 [6] NCCL INFO Trees [0] 7/-1/-1->6->5 [1] 7/-1/-1->6->5 [2] 7/-1/-1->6->5 [3] 7/-1/-1->6->5 [4] 7/-1/-1->6->5 [5] 7/-1/-1->6->5 [6] 7/14/-1->6->-1 [7] -1/-1/-1->6->5 [8] 7/-1/-1->6->5 [9] 7/-1/-1->6->5 [10] 7/-1/-1->6->5 [11] 7/-1/-1->6->5 [12] 7/-1/-1->6->5 [13] 7/-1/-1->6->5 [14] 7/-1/-1->6->14 [15] -1/-1/-1->6->5 cnode7-008:3116465:3117199 [6] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307519:2308205 [0] NCCL INFO comm 0x555564bcae80 rank 8 nRanks 16 nNodes 2 localRanks 8 localRank 0 MNNVL 0 cnode7-009:2307519:2308205 [0] NCCL INFO Trees [0] 9/-1/-1->8->0 [1] -1/-1/-1->8->15 [2] 9/-1/-1->8->15 [3] 9/-1/-1->8->15 [4] 9/-1/-1->8->15 [5] 9/-1/-1->8->15 [6] 9/-1/-1->8->15 [7] 9/-1/-1->8->15 [8] 9/0/-1->8->-1 [9] -1/-1/-1->8->15 [10] 9/-1/-1->8->15 [11] 9/-1/-1->8->15 [12] 9/-1/-1->8->15 [13] 9/-1/-1->8->15 [14] 9/-1/-1->8->15 [15] 9/-1/-1->8->15 cnode7-009:2307519:2308205 [0] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307520:2308272 [1] NCCL INFO comm 0x555560a202e0 rank 9 nRanks 16 nNodes 2 localRanks 8 localRank 1 MNNVL 0 cnode7-009:2307520:2308272 [1] NCCL INFO Trees [0] 10/-1/-1->9->8 [1] 10/-1/-1->9->1 [2] -1/-1/-1->9->8 [3] 10/-1/-1->9->8 [4] 10/-1/-1->9->8 [5] 10/-1/-1->9->8 [6] 10/-1/-1->9->8 [7] 10/-1/-1->9->8 [8] 10/-1/-1->9->8 [9] 10/1/-1->9->-1 [10] -1/-1/-1->9->8 [11] 10/-1/-1->9->8 [12] 10/-1/-1->9->8 [13] 10/-1/-1->9->8 [14] 10/-1/-1->9->8 [15] 10/-1/-1->9->8 cnode7-009:2307520:2308272 [1] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116463:3117200 [4] NCCL INFO comm 0x55557bc1dc20 rank 4 nRanks 16 nNodes 2 localRanks 8 localRank 4 MNNVL 0 cnode7-008:3116463:3117200 [4] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116463:3117200 [4] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116463:3117200 [4] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116463:3117200 [4] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116463:3117200 [4] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116463:3117200 [4] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116463:3117200 [4] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116463:3117200 [4] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116463:3117200 [4] NCCL INFO Trees [0] 5/-1/-1->4->3 [1] 5/-1/-1->4->3 [2] 5/-1/-1->4->3 [3] 5/-1/-1->4->3 [4] 5/12/-1->4->-1 [5] -1/-1/-1->4->3 [6] 5/-1/-1->4->3 [7] 5/-1/-1->4->3 [8] 5/-1/-1->4->3 [9] 5/-1/-1->4->3 [10] 5/-1/-1->4->3 [11] 5/-1/-1->4->3 [12] 5/-1/-1->4->12 [13] -1/-1/-1->4->3 [14] 5/-1/-1->4->3 [15] 5/-1/-1->4->3 cnode7-008:3116463:3117200 [4] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307522:2308212 [3] NCCL INFO comm 0x555580c22f60 rank 11 nRanks 16 nNodes 2 localRanks 8 localRank 3 MNNVL 0 cnode7-009:2307522:2308212 [3] NCCL INFO Trees [0] 12/-1/-1->11->10 [1] 12/-1/-1->11->10 [2] 12/-1/-1->11->10 [3] 12/-1/-1->11->3 [4] -1/-1/-1->11->10 [5] 12/-1/-1->11->10 [6] 12/-1/-1->11->10 [7] 12/-1/-1->11->10 [8] 12/-1/-1->11->10 [9] 12/-1/-1->11->10 [10] 12/-1/-1->11->10 [11] 12/3/-1->11->-1 [12] -1/-1/-1->11->10 [13] 12/-1/-1->11->10 [14] 12/-1/-1->11->10 [15] 12/-1/-1->11->10 cnode7-009:2307522:2308212 [3] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307526:2308203 [7] NCCL INFO comm 0x555560a16e20 rank 15 nRanks 16 nNodes 2 localRanks 8 localRank 7 MNNVL 0 cnode7-009:2307526:2308203 [7] NCCL INFO Trees [0] -1/-1/-1->15->14 [1] 8/-1/-1->15->14 [2] 8/-1/-1->15->14 [3] 8/-1/-1->15->14 [4] 8/-1/-1->15->14 [5] 8/-1/-1->15->14 [6] 8/-1/-1->15->14 [7] 8/-1/-1->15->7 [8] -1/-1/-1->15->14 [9] 8/-1/-1->15->14 [10] 8/-1/-1->15->14 [11] 8/-1/-1->15->14 [12] 8/-1/-1->15->14 [13] 8/-1/-1->15->14 [14] 8/-1/-1->15->14 [15] 8/7/-1->15->-1 cnode7-009:2307526:2308203 [7] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307525:2308209 [6] NCCL INFO Setting affinity for GPU 6 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-009:2307525:2308209 [6] NCCL INFO NVLS multicast support is available on dev 6 cnode7-009:2307525:2308209 [6] NCCL INFO comm 0x555567c1b8e0 rank 14 nRanks 16 nNodes 2 localRanks 8 localRank 6 MNNVL 0 cnode7-009:2307525:2308209 [6] NCCL INFO Trees [0] 15/-1/-1->14->13 [1] 15/-1/-1->14->13 [2] 15/-1/-1->14->13 [3] 15/-1/-1->14->13 [4] 15/-1/-1->14->13 [5] 15/-1/-1->14->13 [6] 15/-1/-1->14->6 [7] -1/-1/-1->14->13 [8] 15/-1/-1->14->13 [9] 15/-1/-1->14->13 [10] 15/-1/-1->14->13 [11] 15/-1/-1->14->13 [12] 15/-1/-1->14->13 [13] 15/-1/-1->14->13 [14] 15/6/-1->14->-1 [15] -1/-1/-1->14->13 cnode7-009:2307525:2308209 [6] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307523:2308206 [4] NCCL INFO Setting affinity for GPU 4 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-009:2307523:2308206 [4] NCCL INFO NVLS multicast support is available on dev 4 cnode7-009:2307523:2308206 [4] NCCL INFO comm 0x55556005a740 rank 12 nRanks 16 nNodes 2 localRanks 8 localRank 4 MNNVL 0 cnode7-009:2307523:2308206 [4] NCCL INFO Trees [0] 13/-1/-1->12->11 [1] 13/-1/-1->12->11 [2] 13/-1/-1->12->11 [3] 13/-1/-1->12->11 [4] 13/-1/-1->12->4 [5] -1/-1/-1->12->11 [6] 13/-1/-1->12->11 [7] 13/-1/-1->12->11 [8] 13/-1/-1->12->11 [9] 13/-1/-1->12->11 [10] 13/-1/-1->12->11 [11] 13/-1/-1->12->11 [12] 13/4/-1->12->-1 [13] -1/-1/-1->12->11 [14] 13/-1/-1->12->11 [15] 13/-1/-1->12->11 cnode7-009:2307523:2308206 [4] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307524:2308273 [5] NCCL INFO comm 0x555560098c20 rank 13 nRanks 16 nNodes 2 localRanks 8 localRank 5 MNNVL 0 cnode7-009:2307524:2308273 [5] NCCL INFO Trees [0] 14/-1/-1->13->12 [1] 14/-1/-1->13->12 [2] 14/-1/-1->13->12 [3] 14/-1/-1->13->12 [4] 14/-1/-1->13->12 [5] 14/-1/-1->13->5 [6] -1/-1/-1->13->12 [7] 14/-1/-1->13->12 [8] 14/-1/-1->13->12 [9] 14/-1/-1->13->12 [10] 14/-1/-1->13->12 [11] 14/-1/-1->13->12 [12] 14/-1/-1->13->12 [13] 14/5/-1->13->-1 [14] -1/-1/-1->13->12 [15] 14/-1/-1->13->12 cnode7-009:2307524:2308273 [5] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116464:3117204 [5] NCCL INFO comm 0x555575421c00 rank 5 nRanks 16 nNodes 2 localRanks 8 localRank 5 MNNVL 0 cnode7-008:3116464:3117204 [5] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116464:3117204 [5] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116464:3117204 [5] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116464:3117204 [5] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116464:3117204 [5] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116464:3117204 [5] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116464:3117204 [5] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116464:3117204 [5] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116464:3117204 [5] NCCL INFO Trees [0] 6/-1/-1->5->4 [1] 6/-1/-1->5->4 [2] 6/-1/-1->5->4 [3] 6/-1/-1->5->4 [4] 6/-1/-1->5->4 [5] 6/13/-1->5->-1 [6] -1/-1/-1->5->4 [7] 6/-1/-1->5->4 [8] 6/-1/-1->5->4 [9] 6/-1/-1->5->4 [10] 6/-1/-1->5->4 [11] 6/-1/-1->5->4 [12] 6/-1/-1->5->4 [13] 6/-1/-1->5->13 [14] -1/-1/-1->5->4 [15] 6/-1/-1->5->4 cnode7-008:3116464:3117204 [5] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116460:3117203 [1] NCCL INFO comm 0x55558101c4b0 rank 1 nRanks 16 nNodes 2 localRanks 8 localRank 1 MNNVL 0 cnode7-008:3116460:3117203 [1] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116460:3117203 [1] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116460:3117203 [1] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116460:3117203 [1] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116460:3117203 [1] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116460:3117203 [1] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116460:3117203 [1] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116460:3117203 [1] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116460:3117203 [1] NCCL INFO Trees [0] 2/-1/-1->1->0 [1] 2/9/-1->1->-1 [2] -1/-1/-1->1->0 [3] 2/-1/-1->1->0 [4] 2/-1/-1->1->0 [5] 2/-1/-1->1->0 [6] 2/-1/-1->1->0 [7] 2/-1/-1->1->0 [8] 2/-1/-1->1->0 [9] 2/-1/-1->1->9 [10] -1/-1/-1->1->0 [11] 2/-1/-1->1->0 [12] 2/-1/-1->1->0 [13] 2/-1/-1->1->0 [14] 2/-1/-1->1->0 [15] 2/-1/-1->1->0 cnode7-008:3116460:3117203 [1] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116461:3117202 [2] NCCL INFO comm 0x5555808662f0 rank 2 nRanks 16 nNodes 2 localRanks 8 localRank 2 MNNVL 0 cnode7-008:3116461:3117202 [2] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116461:3117202 [2] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116461:3117202 [2] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116461:3117202 [2] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116461:3117202 [2] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116461:3117202 [2] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116461:3117202 [2] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116461:3117202 [2] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116461:3117202 [2] NCCL INFO Trees [0] 3/-1/-1->2->1 [1] 3/-1/-1->2->1 [2] 3/10/-1->2->-1 [3] -1/-1/-1->2->1 [4] 3/-1/-1->2->1 [5] 3/-1/-1->2->1 [6] 3/-1/-1->2->1 [7] 3/-1/-1->2->1 [8] 3/-1/-1->2->1 [9] 3/-1/-1->2->1 [10] 3/-1/-1->2->10 [11] -1/-1/-1->2->1 [12] 3/-1/-1->2->1 [13] 3/-1/-1->2->1 [14] 3/-1/-1->2->1 [15] 3/-1/-1->2->1 cnode7-008:3116461:3117202 [2] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116466:3117198 [7] NCCL INFO comm 0x555560c1ca30 rank 7 nRanks 16 nNodes 2 localRanks 8 localRank 7 MNNVL 0 cnode7-008:3116466:3117198 [7] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116466:3117198 [7] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116466:3117198 [7] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116466:3117198 [7] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116466:3117198 [7] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116466:3117198 [7] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116466:3117198 [7] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116466:3117198 [7] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116466:3117198 [7] NCCL INFO Trees [0] -1/-1/-1->7->6 [1] 0/-1/-1->7->6 [2] 0/-1/-1->7->6 [3] 0/-1/-1->7->6 [4] 0/-1/-1->7->6 [5] 0/-1/-1->7->6 [6] 0/-1/-1->7->6 [7] 0/15/-1->7->-1 [8] -1/-1/-1->7->6 [9] 0/-1/-1->7->6 [10] 0/-1/-1->7->6 [11] 0/-1/-1->7->6 [12] 0/-1/-1->7->6 [13] 0/-1/-1->7->6 [14] 0/-1/-1->7->6 [15] 0/-1/-1->7->15 cnode7-008:3116466:3117198 [7] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116462:3117205 [3] NCCL INFO comm 0x5555810234b0 rank 3 nRanks 16 nNodes 2 localRanks 8 localRank 3 MNNVL 0 cnode7-008:3116462:3117205 [3] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116462:3117205 [3] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116462:3117205 [3] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116462:3117205 [3] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116462:3117205 [3] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116462:3117205 [3] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116462:3117205 [3] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116462:3117205 [3] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116462:3117205 [3] NCCL INFO Trees [0] 4/-1/-1->3->2 [1] 4/-1/-1->3->2 [2] 4/-1/-1->3->2 [3] 4/11/-1->3->-1 [4] -1/-1/-1->3->2 [5] 4/-1/-1->3->2 [6] 4/-1/-1->3->2 [7] 4/-1/-1->3->2 [8] 4/-1/-1->3->2 [9] 4/-1/-1->3->2 [10] 4/-1/-1->3->2 [11] 4/-1/-1->3->11 [12] -1/-1/-1->3->2 [13] 4/-1/-1->3->2 [14] 4/-1/-1->3->2 [15] 4/-1/-1->3->2 cnode7-008:3116462:3117205 [3] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116459:3117197 [0] NCCL INFO comm 0x5555647ca110 rank 0 nRanks 16 nNodes 2 localRanks 8 localRank 0 MNNVL 0 cnode7-008:3116459:3117197 [0] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116459:3117197 [0] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116459:3117197 [0] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116459:3117197 [0] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116459:3117197 [0] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116459:3117197 [0] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116459:3117197 [0] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116459:3117197 [0] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 00/16 : 0 7 6 5 4 3 2 1 9 10 11 12 13 14 15 8 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 01/16 : 0 8 15 14 13 12 11 10 9 1 2 3 4 5 6 7 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 02/16 : 0 7 6 5 4 3 11 12 13 14 15 8 9 10 2 1 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 03/16 : 0 1 2 10 9 8 15 14 13 12 11 3 4 5 6 7 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 04/16 : 0 7 6 5 13 14 15 8 9 10 11 12 4 3 2 1 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 05/16 : 0 1 2 3 4 12 11 10 9 8 15 14 13 5 6 7 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 06/16 : 0 7 15 8 9 10 11 12 13 14 6 5 4 3 2 1 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 07/16 : 0 1 2 3 4 5 6 14 13 12 11 10 9 8 15 7 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 08/16 : 0 7 6 5 4 3 2 1 9 10 11 12 13 14 15 8 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 09/16 : 0 8 15 14 13 12 11 10 9 1 2 3 4 5 6 7 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 10/16 : 0 7 6 5 4 3 11 12 13 14 15 8 9 10 2 1 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 11/16 : 0 1 2 10 9 8 15 14 13 12 11 3 4 5 6 7 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 12/16 : 0 7 6 5 13 14 15 8 9 10 11 12 4 3 2 1 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 13/16 : 0 1 2 3 4 12 11 10 9 8 15 14 13 5 6 7 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 14/16 : 0 7 15 8 9 10 11 12 13 14 6 5 4 3 2 1 cnode7-008:3116459:3117197 [0] NCCL INFO Channel 15/16 : 0 1 2 3 4 5 6 14 13 12 11 10 9 8 15 7 cnode7-008:3116459:3117197 [0] NCCL INFO Trees [0] 1/8/-1->0->-1 [1] -1/-1/-1->0->7 [2] 1/-1/-1->0->7 [3] 1/-1/-1->0->7 [4] 1/-1/-1->0->7 [5] 1/-1/-1->0->7 [6] 1/-1/-1->0->7 [7] 1/-1/-1->0->7 [8] 1/-1/-1->0->8 [9] -1/-1/-1->0->7 [10] 1/-1/-1->0->7 [11] 1/-1/-1->0->7 [12] 1/-1/-1->0->7 [13] 1/-1/-1->0->7 [14] 1/-1/-1->0->7 [15] 1/-1/-1->0->7 cnode7-008:3116459:3117197 [0] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307525:2308209 [6] NCCL INFO Channel 00/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 02/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 00/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 02/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 06/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 08/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 10/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 14/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 01/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 03/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 05/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 07/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 09/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 11/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 13/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 15/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 00/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 04/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 06/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 08/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 12/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 14/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 03/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 11/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 02/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 10/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 00/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 02/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 04/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 06/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 08/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 10/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 12/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 14/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 02/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 10/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 03/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 11/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 00/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 02/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 04/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 06/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 08/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 10/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 12/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 14/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 00/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 08/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 01/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 09/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 01/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 03/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 05/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 09/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 11/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 13/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 06/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 14/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 07/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 15/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 01/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 03/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 07/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 09/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 11/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 15/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 04/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 12/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 05/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 13/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 02/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 04/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 06/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 10/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 12/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 14/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 01/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 03/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 05/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 07/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 09/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 11/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 13/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 15/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 01/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 03/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 05/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 07/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 09/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 11/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 13/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 15/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 01/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 05/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 07/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 09/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 13/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 15/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 02/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 10/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 03/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 11/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 00/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 02/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 04/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 06/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 08/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 10/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 12/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 14/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 01/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 03/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 05/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 07/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 09/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 11/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 13/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 15/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 03/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 11/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 02/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 10/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 00/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 04/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 06/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 08/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 12/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 14/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 04/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 08/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 10/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 12/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 07/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 15/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 06/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 14/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 01/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 03/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 05/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 07/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 09/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 11/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 13/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 05/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 13/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 04/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 12/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 01/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 03/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 05/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 07/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 09/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 11/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 13/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 15/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 05/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 13/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 04/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 12/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 00/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 02/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 06/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 08/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 10/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 14/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 03/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 05/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 07/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 11/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 13/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 15/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 00/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 02/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 04/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 06/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 08/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 10/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 12/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 14/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 00/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 08/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 01/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 09/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 01/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 03/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 05/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 07/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 09/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 11/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 13/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 15/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 01/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 05/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 07/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 09/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 13/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 15/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 06/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 14/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 07/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 15/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 00/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 02/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 04/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 06/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 08/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 10/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 12/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 14/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 01/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 03/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 03/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 05/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 07/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 11/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 13/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 15/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 00/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 02/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 04/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 06/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 08/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 10/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 12/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 14/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 00/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 02/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 04/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 06/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 08/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 10/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 12/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 14/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 01/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 09/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 00/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 08/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 02/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 04/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 06/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 10/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 12/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 14/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 00/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 02/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 04/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 06/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 08/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 10/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 12/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 14/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 04/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 12/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 05/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 13/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 01/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 03/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 07/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 09/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 11/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 15/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 01/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 09/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 00/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 08/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 07/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 15/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 06/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 14/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 01/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 03/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 05/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 07/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 09/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 11/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 13/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 15/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 15/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 05/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 09/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 11/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 13/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 00/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 02/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 04/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 08/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 10/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 12/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Connected all rings cnode7-009:2307521:2308204 [2] NCCL INFO Connected all rings cnode7-009:2307521:2308204 [2] NCCL INFO Channel 01/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 02/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Connected all rings cnode7-008:3116463:3117200 [4] NCCL INFO Connected all rings cnode7-008:3116463:3117200 [4] NCCL INFO Channel 00/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 02/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 04/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 06/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 08/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 10/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 12/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Connected all rings cnode7-008:3116465:3117199 [6] NCCL INFO Channel 00/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 02/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 04/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 06/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 08/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 10/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Connected all rings cnode7-008:3116460:3117203 [1] NCCL INFO Channel 00/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 04/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 06/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 08/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 12/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Connected all rings cnode7-008:3116461:3117202 [2] NCCL INFO Channel 00/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 02/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 04/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 06/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 08/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 10/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 12/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 14/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 02/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 10/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Connected all rings cnode7-009:2307524:2308273 [5] NCCL INFO Channel 01/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 03/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 05/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 07/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 09/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Connected all rings cnode7-009:2307519:2308205 [0] NCCL INFO Channel 00/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 03/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 05/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 07/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 08/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 11/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 13/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 15/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Connected all rings cnode7-008:3116466:3117198 [7] NCCL INFO Channel 07/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 15/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 00/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 02/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 06/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 08/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 10/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 14/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 03/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 11/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 01/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 02/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 05/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 07/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 09/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 10/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 13/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3117205 [3] NCCL INFO Channel 15/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Connected all rings cnode7-009:2307525:2308209 [6] NCCL INFO Channel 01/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 03/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 05/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 06/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 09/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 11/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 13/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 14/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 06/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 14/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 00/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 02/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 04/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 08/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 10/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2308209 [6] NCCL INFO Channel 12/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Connected all rings cnode7-009:2307523:2308206 [4] NCCL INFO Channel 01/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 03/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 04/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 07/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 09/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 11/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 12/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 15/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 04/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 12/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 00/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 02/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 06/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 08/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 10/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2308206 [4] NCCL INFO Channel 14/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Connected all rings cnode7-008:3116464:3117204 [5] NCCL INFO Channel 00/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 02/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 04/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 08/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 10/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 12/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 05/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 13/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 01/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 03/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 04/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 07/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 09/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 11/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 12/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3117204 [5] NCCL INFO Channel 15/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Connected all rings cnode7-008:3116459:3117197 [0] NCCL INFO Channel 00/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 02/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 04/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 06/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 08/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 10/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 12/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 14/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 01/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 03/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 05/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 07/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 09/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 11/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 13/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 15/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3117197 [0] NCCL INFO Channel 00/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 08/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 05/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 07/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 09/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 10/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 13/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 15/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 02/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 10/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 00/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 04/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 06/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 08/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 12/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2308204 [2] NCCL INFO Channel 14/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 01/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 03/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 05/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 07/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 09/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 11/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 13/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 15/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 03/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 11/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 00/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 02/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 04/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 06/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 08/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 10/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 12/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2308212 [3] NCCL INFO Channel 14/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Connected all rings cnode7-009:2307526:2308203 [7] NCCL INFO Channel 07/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 15/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 01/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 03/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 05/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 07/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 09/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 11/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 13/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 15/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 00/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 02/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 04/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 06/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 08/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 10/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 12/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2308203 [7] NCCL INFO Channel 14/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Connected all rings cnode7-009:2307520:2308272 [1] NCCL INFO Channel 01/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 03/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 05/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 07/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 09/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 11/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 13/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 15/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 01/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 09/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 00/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 02/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 04/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 06/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 08/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 10/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 12/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2308272 [1] NCCL INFO Channel 14/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 14/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 04/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 12/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 01/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 03/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 05/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 07/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 09/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 11/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 13/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3117200 [4] NCCL INFO Channel 15/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 12/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 14/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 06/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 14/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 01/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 03/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 05/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 07/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 09/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 11/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 13/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3117199 [6] NCCL INFO Channel 15/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 14/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 01/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 09/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 00/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 03/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 05/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 07/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 08/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 11/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 13/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3117203 [1] NCCL INFO Channel 15/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 01/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 03/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 05/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 07/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 09/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 11/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 13/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Channel 15/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 11/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 13/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 15/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 05/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 13/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 00/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 02/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 04/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 06/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 08/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 10/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 12/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2308273 [5] NCCL INFO Channel 14/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 02/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 04/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 06/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 10/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 12/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 14/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2308205 [0] NCCL INFO Channel 00/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 08/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 02/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 04/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 06/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 10/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 12/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 14/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 01/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 03/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 05/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 06/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 09/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 11/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 13/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3117198 [7] NCCL INFO Channel 14/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116461:3117202 [2] NCCL INFO Connected all trees cnode7-008:3116459:3117197 [0] NCCL INFO Connected all trees cnode7-008:3116460:3117203 [1] NCCL INFO Connected all trees cnode7-009:2307521:2308204 [2] NCCL INFO Connected all trees cnode7-009:2307519:2308205 [0] NCCL INFO Connected all trees cnode7-009:2307520:2308272 [1] NCCL INFO Connected all trees cnode7-009:2307522:2308212 [3] NCCL INFO Connected all trees cnode7-009:2307526:2308203 [7] NCCL INFO Connected all trees cnode7-009:2307524:2308273 [5] NCCL INFO Connected all trees cnode7-009:2307525:2308209 [6] NCCL INFO Connected all trees cnode7-009:2307523:2308206 [4] NCCL INFO Connected all trees cnode7-008:3116463:3117200 [4] NCCL INFO Connected all trees cnode7-008:3116464:3117204 [5] NCCL INFO Connected all trees cnode7-008:3116466:3117198 [7] NCCL INFO Connected all trees cnode7-008:3116462:3117205 [3] NCCL INFO Connected all trees cnode7-008:3116465:3117199 [6] NCCL INFO Connected all trees cnode7-009:2307525:2308209 [6] NCCL INFO NVLS comm 0x555567c1b8e0 headRank 6 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307523:2308206 [4] NCCL INFO NVLS comm 0x55556005a740 headRank 4 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116464:3117204 [5] NCCL INFO NVLS comm 0x555575421c00 headRank 5 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116459:3117197 [0] NCCL INFO NVLS comm 0x5555647ca110 headRank 0 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307521:2308204 [2] NCCL INFO NVLS comm 0x5555810227a0 headRank 2 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307522:2308212 [3] NCCL INFO NVLS comm 0x555580c22f60 headRank 3 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307526:2308203 [7] NCCL INFO NVLS comm 0x555560a16e20 headRank 7 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307520:2308272 [1] NCCL INFO NVLS comm 0x555560a202e0 headRank 1 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116463:3117200 [4] NCCL INFO NVLS comm 0x55557bc1dc20 headRank 4 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116465:3117199 [6] NCCL INFO NVLS comm 0x555560460a80 headRank 6 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116460:3117203 [1] NCCL INFO NVLS comm 0x55558101c4b0 headRank 1 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116461:3117202 [2] NCCL INFO NVLS comm 0x5555808662f0 headRank 2 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307524:2308273 [5] NCCL INFO NVLS comm 0x555560098c20 headRank 5 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307519:2308205 [0] NCCL INFO NVLS comm 0x555564bcae80 headRank 0 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116466:3117198 [7] NCCL INFO NVLS comm 0x555560c1ca30 headRank 7 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116462:3117205 [3] NCCL INFO NVLS comm 0x5555810234b0 headRank 3 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116462:3117205 [3] NCCL INFO Channel 00/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 01/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 02/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 04/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 05/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 06/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 07/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 08/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 00/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 01/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 02/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 03/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 04/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 05/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 08/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 09/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 00/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 01/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 02/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 03/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 06/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 07/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 08/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 09/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 00/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 01/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 02/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 03/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 04/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 06/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 07/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 08/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 01/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 02/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 03/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 04/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 05/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 06/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 07/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 09/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 00/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 01/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 04/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 05/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 06/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 07/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 08/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 09/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 00/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 01/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 04/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 05/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 06/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 07/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 08/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 09/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 00/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 01/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 02/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 03/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 04/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 05/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 08/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 09/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 02/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 03/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 04/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 05/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 06/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 07/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 10/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 11/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 00/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 01/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 02/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 03/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 05/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 06/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 07/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 08/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 00/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 01/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 02/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 03/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 04/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 05/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 07/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 08/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 00/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 01/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 02/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 03/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 06/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 07/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 08/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 09/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 10/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 11/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 14/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 15/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 00/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 01/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 02/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 03/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 04/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 06/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 07/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 08/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 09/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 10/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 11/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 12/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 14/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2308273 [5] NCCL INFO Channel 15/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 02/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 03/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 04/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 05/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 06/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 07/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 10/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 11/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 12/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 13/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 14/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 15/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 01/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 02/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 03/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 04/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 05/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 06/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 07/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 09/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 10/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 11/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 12/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 13/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 14/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2308205 [0] NCCL INFO Channel 15/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 00/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 02/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 03/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 04/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 05/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 06/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 07/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 08/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 00/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 01/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 03/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 04/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 05/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 06/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 07/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 08/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 00/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 01/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 02/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 03/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 04/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 05/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 06/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 08/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 09/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 10/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 12/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 13/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 14/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 15/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 00/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 01/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 04/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 05/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 06/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 07/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 08/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 09/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 12/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 13/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 14/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Channel 15/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 10/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 11/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 12/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 13/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 00/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 01/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 02/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 03/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 04/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 05/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 07/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 08/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 09/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 10/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 11/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 12/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 13/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2308209 [6] NCCL INFO Channel 15/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 10/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 11/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 14/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 15/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 00/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 01/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 02/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 03/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 05/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 06/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 07/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 08/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 09/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 10/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 11/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 13/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 14/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2308206 [4] NCCL INFO Channel 15/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 09/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 10/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 11/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 12/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 14/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 15/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 00/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 01/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 02/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 03/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 06/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 07/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 08/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 09/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 10/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 11/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 14/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3117204 [5] NCCL INFO Channel 15/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 10/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 11/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 12/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 13/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 14/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 15/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 02/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 03/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 04/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 05/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 06/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 07/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 10/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 11/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 12/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 13/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 14/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3117197 [0] NCCL INFO Channel 15/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 12/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 13/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 14/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 15/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 00/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 01/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 03/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 04/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 05/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 06/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 07/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 08/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 09/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 11/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 12/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 13/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 14/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2308204 [2] NCCL INFO Channel 15/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 12/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 13/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 14/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 15/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 00/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 01/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 02/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 04/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 05/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 06/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 07/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 08/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 09/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 10/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 12/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 13/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 14/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2308212 [3] NCCL INFO Channel 15/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 10/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 11/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 12/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 13/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 00/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 01/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 02/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 03/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 04/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 05/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 06/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 08/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 09/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 10/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 11/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 12/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 13/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2308203 [7] NCCL INFO Channel 14/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 12/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 13/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 14/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 15/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 00/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 02/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 03/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 04/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 05/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 06/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 07/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 08/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 10/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 11/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 12/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 13/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 14/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2308272 [1] NCCL INFO Channel 15/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 09/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 10/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 11/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 13/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 14/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 15/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 00/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 01/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 02/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 03/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 06/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 07/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 08/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 09/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 10/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 11/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 14/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3117200 [4] NCCL INFO Channel 15/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 09/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 10/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 11/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 12/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 13/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 15/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 00/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 01/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 02/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 03/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 04/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 05/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 08/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 09/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 10/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 11/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 12/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3117199 [6] NCCL INFO Channel 13/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 10/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 11/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 12/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 13/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 14/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 15/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 02/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 03/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 04/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 05/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 06/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 07/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 10/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 11/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 12/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 13/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 14/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3117203 [1] NCCL INFO Channel 15/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 09/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 11/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 12/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 13/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 14/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 15/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 00/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 01/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 04/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 05/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 06/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 07/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 08/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 09/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 12/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 13/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 14/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3117202 [2] NCCL INFO Channel 15/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 09/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 10/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 11/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 12/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 13/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 14/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 00/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 01/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 02/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 03/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 04/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 05/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 08/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 09/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 10/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 11/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 12/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3117198 [7] NCCL INFO Channel 13/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116462:3117205 [3] NCCL INFO Connected NVLS tree cnode7-008:3116462:3117205 [3] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116462:3117205 [3] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116464:3117204 [5] NCCL INFO Connected NVLS tree cnode7-008:3116464:3117204 [5] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116464:3117204 [5] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116460:3117203 [1] NCCL INFO Connected NVLS tree cnode7-008:3116460:3117203 [1] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116460:3117203 [1] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307525:2308209 [6] NCCL INFO Connected NVLS tree cnode7-009:2307525:2308209 [6] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307525:2308209 [6] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116459:3117197 [0] NCCL INFO Connected NVLS tree cnode7-008:3116459:3117197 [0] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116459:3117197 [0] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307521:2308204 [2] NCCL INFO Connected NVLS tree cnode7-009:2307521:2308204 [2] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307521:2308204 [2] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307522:2308212 [3] NCCL INFO Connected NVLS tree cnode7-009:2307522:2308212 [3] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307522:2308212 [3] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307526:2308203 [7] NCCL INFO Connected NVLS tree cnode7-009:2307526:2308203 [7] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307526:2308203 [7] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307520:2308272 [1] NCCL INFO Connected NVLS tree cnode7-009:2307520:2308272 [1] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307520:2308272 [1] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116461:3117202 [2] NCCL INFO Connected NVLS tree cnode7-008:3116461:3117202 [2] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116461:3117202 [2] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116465:3117199 [6] NCCL INFO Connected NVLS tree cnode7-008:3116465:3117199 [6] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116465:3117199 [6] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116466:3117198 [7] NCCL INFO Connected NVLS tree cnode7-008:3116466:3117198 [7] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116466:3117198 [7] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307524:2308273 [5] NCCL INFO Connected NVLS tree cnode7-009:2307524:2308273 [5] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307524:2308273 [5] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307519:2308205 [0] NCCL INFO Connected NVLS tree cnode7-009:2307519:2308205 [0] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307519:2308205 [0] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116463:3117200 [4] NCCL INFO Connected NVLS tree cnode7-008:3116463:3117200 [4] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116463:3117200 [4] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116465:3117199 [6] NCCL INFO comm 0x555560460a80 rank 6 nranks 16 cudaDev 6 nvmlDev 6 busId d1000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-008:3116459:3117197 [0] NCCL INFO comm 0x5555647ca110 rank 0 nranks 16 cudaDev 0 nvmlDev 0 busId 1b000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-008:3116463:3117200 [4] NCCL INFO comm 0x55557bc1dc20 rank 4 nranks 16 cudaDev 4 nvmlDev 4 busId 9d000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-008:3116464:3117204 [5] NCCL INFO comm 0x555575421c00 rank 5 nranks 16 cudaDev 5 nvmlDev 5 busId c3000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-008:3116460:3117203 [1] NCCL INFO comm 0x55558101c4b0 rank 1 nranks 16 cudaDev 1 nvmlDev 1 busId 43000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-008:3116466:3117198 [7] NCCL INFO comm 0x555560c1ca30 rank 7 nranks 16 cudaDev 7 nvmlDev 7 busId df000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-009:2307523:2308206 [4] NCCL INFO Connected NVLS tree cnode7-009:2307523:2308206 [4] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307523:2308206 [4] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116461:3117202 [2] NCCL INFO comm 0x5555808662f0 rank 2 nranks 16 cudaDev 2 nvmlDev 2 busId 52000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-008:3116462:3117205 [3] NCCL INFO comm 0x5555810234b0 rank 3 nranks 16 cudaDev 3 nvmlDev 3 busId 61000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-009:2307524:2308273 [5] NCCL INFO comm 0x555560098c20 rank 13 nranks 16 cudaDev 5 nvmlDev 5 busId c3000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-009:2307519:2308205 [0] NCCL INFO comm 0x555564bcae80 rank 8 nranks 16 cudaDev 0 nvmlDev 0 busId 1b000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-009:2307520:2308272 [1] NCCL INFO comm 0x555560a202e0 rank 9 nranks 16 cudaDev 1 nvmlDev 1 busId 43000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-009:2307525:2308209 [6] NCCL INFO comm 0x555567c1b8e0 rank 14 nranks 16 cudaDev 6 nvmlDev 6 busId d1000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-009:2307522:2308212 [3] NCCL INFO comm 0x555580c22f60 rank 11 nranks 16 cudaDev 3 nvmlDev 3 busId 61000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-009:2307521:2308204 [2] NCCL INFO comm 0x5555810227a0 rank 10 nranks 16 cudaDev 2 nvmlDev 2 busId 52000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-009:2307526:2308203 [7] NCCL INFO comm 0x555560a16e20 rank 15 nranks 16 cudaDev 7 nvmlDev 7 busId df000 commId 0x821cf94b50e14dd2 - Init COMPLETE cnode7-009:2307523:2308206 [4] NCCL INFO comm 0x55556005a740 rank 12 nranks 16 cudaDev 4 nvmlDev 4 busId 9d000 commId 0x821cf94b50e14dd2 - Init COMPLETE [2024-11-29 11:10:03,164] [INFO] [logging.py:129:log_dist] [Rank 0] DeepSpeed Flops Profiler Enabled: False [2024-11-29 11:10:03,174] [INFO] [logging.py:129:log_dist] [Rank 0] Using DeepSpeed Optimizer param name adam as basic optimizer [2024-11-29 11:10:03,174] [INFO] [logging.py:129:log_dist] [Rank 0] Removing param_group that has no 'params' in the basic Optimizer [2024-11-29 11:10:03,208] [INFO] [logging.py:129:log_dist] [Rank 0] DeepSpeed Basic Optimizer = FusedAdam [2024-11-29 11:10:03,208] [INFO] [utils.py:59:is_zero_supported_optimizer] Checking ZeRO support for optimizer=FusedAdam type= [2024-11-29 11:10:03,208] [INFO] [logging.py:129:log_dist] [Rank 0] Creating torch.bfloat16 ZeRO stage 2 optimizer [2024-11-29 11:10:03,208] [INFO] [stage_1_and_2.py:149:__init__] Reduce bucket size 500000000 [2024-11-29 11:10:03,208] [INFO] [stage_1_and_2.py:150:__init__] Allgather bucket size 500000000 [2024-11-29 11:10:03,208] [INFO] [stage_1_and_2.py:151:__init__] CPU Offload: False [2024-11-29 11:10:03,208] [INFO] [stage_1_and_2.py:152:__init__] Round robin gradient partitioning: False [2024-11-29 11:10:04,761] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:04,762] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:04,761] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:04,761] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:04,762] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) [2024-11-29 11:09:51,977] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:09:51,978] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:04,769] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) [2024-11-29 11:09:51,984] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) [2024-11-29 11:09:51,986] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) [2024-11-29 11:10:04,778] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) [2024-11-29 11:09:52,005] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) [2024-11-29 11:09:52,139] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:09:52,139] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) [2024-11-29 11:10:05,065] [INFO] [utils.py:781:see_memory_usage] Before initializing optimizer states [2024-11-29 11:10:05,068] [INFO] [utils.py:782:see_memory_usage] MA 1.37 GB Max_MA 1.45 GB CA 1.45 GB Max_CA 1 GB [2024-11-29 11:10:05,068] [INFO] [utils.py:789:see_memory_usage] CPU Virtual Memory: used = 72.87 GB, percent = 3.6% [2024-11-29 11:09:52,314] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) [2024-11-29 11:10:05,255] [INFO] [utils.py:781:see_memory_usage] After initializing optimizer states [2024-11-29 11:10:05,256] [INFO] [utils.py:782:see_memory_usage] MA 1.37 GB Max_MA 1.52 GB CA 1.6 GB Max_CA 2 GB [2024-11-29 11:10:05,256] [INFO] [utils.py:789:see_memory_usage] CPU Virtual Memory: used = 75.0 GB, percent = 3.7% [2024-11-29 11:10:05,256] [INFO] [stage_1_and_2.py:544:__init__] optimizer state initialized [2024-11-29 11:10:05,417] [INFO] [utils.py:781:see_memory_usage] After initializing ZeRO optimizer [2024-11-29 11:10:05,417] [INFO] [utils.py:782:see_memory_usage] MA 1.37 GB Max_MA 1.37 GB CA 1.6 GB Max_CA 2 GB [2024-11-29 11:10:05,417] [INFO] [utils.py:789:see_memory_usage] CPU Virtual Memory: used = 75.28 GB, percent = 3.7% [2024-11-29 11:10:05,419] [INFO] [logging.py:129:log_dist] [Rank 0] DeepSpeed Final Optimizer = DeepSpeedZeroOptimizer [2024-11-29 11:10:05,419] [INFO] [logging.py:129:log_dist] [Rank 0] DeepSpeed using client callable to create LR scheduler [2024-11-29 11:10:05,419] [INFO] [logging.py:129:log_dist] [Rank 0] DeepSpeed LR Scheduler = PiecewiseLinearWarmupLR(warmup_steps_list=[0, 30000, 60000], warmup_lr_list=[0.0, 5e-05, 0.0002]) [2024-11-29 11:10:05,419] [INFO] [logging.py:129:log_dist] [Rank 0] step=0, skipped=0, lr=[np.float64(1.6666666666666667e-09)], mom=[[0.9, 0.98]] [2024-11-29 11:10:05,421] [INFO] [config.py:999:print] DeepSpeedEngine configuration: [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] activation_checkpointing_config { "partition_activations": false, "contiguous_memory_optimization": false, "cpu_checkpointing": false, "number_checkpoints": null, "synchronize_checkpoint_boundary": false, "profile": false } [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] aio_config ................... {'block_size': 1048576, 'queue_depth': 8, 'thread_count': 1, 'single_submit': False, 'overlap_events': True, 'use_gds': False} [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] amp_enabled .................. False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] amp_params ................... False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] autotuning_config ............ { "enabled": false, "start_step": null, "end_step": null, "metric_path": null, "arg_mappings": null, "metric": "throughput", "model_info": null, "results_dir": "autotuning_results", "exps_dir": "autotuning_exps", "overwrite": true, "fast": true, "start_profile_step": 3, "end_profile_step": 5, "tuner_type": "gridsearch", "tuner_early_stopping": 5, "tuner_num_trials": 50, "model_info_path": null, "mp_size": 1, "max_train_batch_size": null, "min_train_batch_size": 1, "max_train_micro_batch_size_per_gpu": 1.024000e+03, "min_train_micro_batch_size_per_gpu": 1, "num_tuning_micro_batch_sizes": 3 } [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] bfloat16_enabled ............. True [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] bfloat16_immediate_grad_update False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] checkpoint_parallel_write_pipeline False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] checkpoint_tag_validation_enabled True [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] checkpoint_tag_validation_fail False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] comms_config ................. [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] communication_data_type ...... None [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] compression_config ........... {'weight_quantization': {'shared_parameters': {'enabled': False, 'quantizer_kernel': False, 'schedule_offset': 0, 'quantize_groups': 1, 'quantize_verbose': False, 'quantization_type': 'symmetric', 'quantize_weight_in_forward': False, 'rounding': 'nearest', 'fp16_mixed_quantize': False, 'quantize_change_ratio': 0.001}, 'different_groups': {}}, 'activation_quantization': {'shared_parameters': {'enabled': False, 'quantization_type': 'symmetric', 'range_calibration': 'dynamic', 'schedule_offset': 1000}, 'different_groups': {}}, 'sparse_pruning': {'shared_parameters': {'enabled': False, 'method': 'l1', 'schedule_offset': 1000}, 'different_groups': {}}, 'row_pruning': {'shared_parameters': {'enabled': False, 'method': 'l1', 'schedule_offset': 1000}, 'different_groups': {}}, 'head_pruning': {'shared_parameters': {'enabled': False, 'method': 'topk', 'schedule_offset': 1000}, 'different_groups': {}}, 'channel_pruning': {'shared_parameters': {'enabled': False, 'method': 'l1', 'schedule_offset': 1000}, 'different_groups': {}}, 'layer_reduction': {'enabled': False}} [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] curriculum_enabled_legacy .... False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] curriculum_params_legacy ..... False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] data_efficiency_config ....... {'enabled': False, 'seed': 1234, 'data_sampling': {'enabled': False, 'num_epochs': 1000, 'num_workers': 0, 'curriculum_learning': {'enabled': False}}, 'data_routing': {'enabled': False, 'random_ltd': {'enabled': False, 'layer_token_lr_schedule': {'enabled': False}}}} [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] data_efficiency_enabled ...... False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] dataloader_drop_last ......... False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] disable_allgather ............ False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] dump_state ................... False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] dynamic_loss_scale_args ...... None [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] eigenvalue_enabled ........... False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] eigenvalue_gas_boundary_resolution 1 [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] eigenvalue_layer_name ........ bert.encoder.layer [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] eigenvalue_layer_num ......... 0 [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] eigenvalue_max_iter .......... 100 [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] eigenvalue_stability ......... 1e-06 [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] eigenvalue_tol ............... 0.01 [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] eigenvalue_verbose ........... False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] elasticity_enabled ........... False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] flops_profiler_config ........ { "enabled": false, "recompute_fwd_factor": 0.0, "profile_step": 1, "module_depth": -1, "top_modules": 1, "detailed": true, "output_file": null } [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] fp16_auto_cast ............... None [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] fp16_enabled ................. False [2024-11-29 11:10:05,421] [INFO] [config.py:1003:print] fp16_master_weights_and_gradients False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] global_rank .................. 0 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] grad_accum_dtype ............. None [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] gradient_accumulation_steps .. 1 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] gradient_clipping ............ 5.0 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] gradient_predivide_factor .... 1.0 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] graph_harvesting ............. False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] hybrid_engine ................ enabled=False max_out_tokens=512 inference_tp_size=1 release_inference_cache=False pin_parameters=True tp_gather_partition_size=8 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] initial_dynamic_scale ........ 1 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] load_universal_checkpoint .... False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] loss_scale ................... 1.0 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] memory_breakdown ............. False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] mics_hierarchial_params_gather False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] mics_shard_size .............. -1 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] monitor_config ............... tensorboard=TensorBoardConfig(enabled=False, output_path='', job_name='DeepSpeedJobName') comet=CometConfig(enabled=False, samples_log_interval=100, project=None, workspace=None, api_key=None, experiment_name=None, experiment_key=None, online=None, mode=None) wandb=WandbConfig(enabled=False, group=None, team=None, project='deepspeed') csv_monitor=CSVConfig(enabled=False, output_path='', job_name='DeepSpeedJobName') [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] nebula_config ................ { "enabled": false, "persistent_storage_path": null, "persistent_time_interval": 100, "num_of_version_in_retention": 2, "enable_nebula_load": true, "load_path": null } [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] optimizer_legacy_fusion ...... False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] optimizer_name ............... adam [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] optimizer_params ............. {'lr': 0.0002, 'betas': [0.9, 0.98], 'eps': 1e-06, 'weight_decay': 0.0} [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] pipeline ..................... {'stages': 'auto', 'partition': 'best', 'seed_layers': False, 'activation_checkpoint_interval': 0, 'pipe_partitioned': True, 'grad_partitioned': True} [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] pld_enabled .................. False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] pld_params ................... False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] prescale_gradients ........... False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] scheduler_name ............... None [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] scheduler_params ............. None [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] seq_parallel_communication_data_type torch.float32 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] sparse_attention ............. None [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] sparse_gradients_enabled ..... False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] steps_per_print .............. 1000 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] timers_config ................ enabled=True synchronized=True [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] train_batch_size ............. 16 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] train_micro_batch_size_per_gpu 1 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] use_data_before_expert_parallel_ False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] use_node_local_storage ....... False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] wall_clock_breakdown ......... False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] weight_quantization_config ... None [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] world_size ................... 16 [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] zero_allow_untested_optimizer False [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] zero_config .................. stage=2 contiguous_gradients=True reduce_scatter=True reduce_bucket_size=500000000 use_multi_rank_bucket_allreduce=True allgather_partitions=True allgather_bucket_size=500000000 overlap_comm=True load_from_fp32_weights=True elastic_checkpoint=False offload_param=None offload_optimizer=None sub_group_size=1000000000 cpu_offload_param=None cpu_offload_use_pin_memory=None cpu_offload=None prefetch_bucket_size=50000000 param_persistence_threshold=100000 model_persistence_threshold=9223372036854775807 max_live_parameters=1000000000 max_reuse_distance=1000000000 gather_16bit_weights_on_model_save=False use_all_reduce_for_fetch_params=False stage3_gather_fp16_weights_on_model_save=False ignore_unused_parameters=True legacy_stage1=False round_robin_gradients=False zero_hpz_partition_size=1 zero_quantized_weights=False zero_quantized_nontrainable_weights=False zero_quantized_gradients=False mics_shard_size=-1 mics_hierarchical_params_gather=False memory_efficient_linear=True pipeline_loading_checkpoint=False override_module_apply=True [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] zero_enabled ................. True [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] zero_force_ds_cpu_optimizer .. True [2024-11-29 11:10:05,422] [INFO] [config.py:1003:print] zero_optimization_stage ...... 2 [2024-11-29 11:10:05,422] [INFO] [config.py:989:print_user_config] json = { "train_micro_batch_size_per_gpu": 1, "gradient_accumulation_steps": 1, "gradient_clipping": 5.0, "bf16": { "enabled": true }, "zero_optimization": { "stage": 2, "contiguous_gradients": true, "overlap_comm": true, "reduce_scatter": true, "reduce_bucket_size": 5.000000e+08, "allgather_bucket_size": 5.000000e+08 }, "optimizer": { "type": "Adam", "params": { "lr": 0.0002, "betas": [0.9, 0.98], "eps": 1e-06, "weight_decay": 0.0 } }, "wall_clock_breakdown": false, "steps_per_print": 1000 } [cnode7-008:0/16] 2024-11-29 11:10:05,424 (deepspeed_trainer:75) INFO: Resume training from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8 [2024-11-29 11:10:05,426] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... /mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/checkpoint_engine/torch_checkpoint_engine.py:28: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature. partition = torch.load(path, map_location=map_location) [2024-11-29 11:10:05,919] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:05,919] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:05,919] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:05,919] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:05,919] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:05,919] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:05,919] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:05,970] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:05,996] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:05,998] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:05,999] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:05,999] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:06,000] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:06,000] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:09:53,236] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:53,236] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:53,236] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:53,236] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:53,236] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:53,236] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:53,236] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:53,236] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:53,294] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:09:53,299] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:09:53,310] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:09:53,310] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:09:53,311] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:09:53,312] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:09:53,313] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:09:53,313] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:06,126] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:06,165] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt... [2024-11-29 11:10:06,641] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:06,659] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:06,659] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:06,659] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:06,659] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:06,659] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:06,661] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:54,014] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:54,014] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:54,014] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:54,015] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:54,015] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:54,015] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:54,015] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:09:54,026] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:06,839] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/mp_rank_00_model_states.pt. [2024-11-29 11:10:06,924] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt... [2024-11-29 11:10:06,924] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt... [2024-11-29 11:10:06,924] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt... [2024-11-29 11:10:06,924] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt... [2024-11-29 11:10:06,924] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt... [2024-11-29 11:10:06,924] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt... [2024-11-29 11:10:06,926] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt... [2024-11-29 11:09:54,252] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_13_mp_rank_00_optim_states.pt... [2024-11-29 11:09:54,252] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_8_mp_rank_00_optim_states.pt... [2024-11-29 11:09:54,253] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_15_mp_rank_00_optim_states.pt... [2024-11-29 11:09:54,262] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_10_mp_rank_00_optim_states.pt... [2024-11-29 11:10:07,055] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt... [2024-11-29 11:09:54,305] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_9_mp_rank_00_optim_states.pt... [2024-11-29 11:09:54,306] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_14_mp_rank_00_optim_states.pt... [2024-11-29 11:09:54,306] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_11_mp_rank_00_optim_states.pt... [2024-11-29 11:09:54,308] [INFO] [torch_checkpoint_engine.py:27:load] [Torch] Loading checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_12_mp_rank_00_optim_states.pt... [2024-11-29 11:10:07,391] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt. [2024-11-29 11:10:07,391] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 2 [2024-11-29 11:10:07,398] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt. [2024-11-29 11:10:07,398] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 7 [2024-11-29 11:10:07,426] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt. [2024-11-29 11:10:07,426] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 3 [2024-11-29 11:10:07,430] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt. [2024-11-29 11:10:07,430] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 1 [2024-11-29 11:10:07,440] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt. [2024-11-29 11:10:07,440] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 6 [2024-11-29 11:10:07,465] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 2 [2024-11-29 11:10:07,475] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 7 [2024-11-29 11:10:07,480] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 3 [2024-11-29 11:10:07,491] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 1 [2024-11-29 11:10:07,503] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt. [2024-11-29 11:10:07,504] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 4 [2024-11-29 11:10:07,505] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 6 [2024-11-29 11:10:07,541] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt. [2024-11-29 11:10:07,542] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 5 [2024-11-29 11:10:07,568] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 4 [2024-11-29 11:09:54,779] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_9_mp_rank_00_optim_states.pt. [2024-11-29 11:09:54,779] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 9 [2024-11-29 11:09:54,785] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_15_mp_rank_00_optim_states.pt. [2024-11-29 11:09:54,785] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 15 [2024-11-29 11:09:54,796] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_13_mp_rank_00_optim_states.pt. [2024-11-29 11:09:54,797] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 13 [2024-11-29 11:10:07,600] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 5 [2024-11-29 11:09:54,821] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_14_mp_rank_00_optim_states.pt. [2024-11-29 11:09:54,822] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 14 [2024-11-29 11:09:54,834] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_10_mp_rank_00_optim_states.pt. [2024-11-29 11:09:54,834] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 10 [2024-11-29 11:09:54,836] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_8_mp_rank_00_optim_states.pt. [2024-11-29 11:09:54,837] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 8 [2024-11-29 11:09:54,839] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 9 [2024-11-29 11:10:07,633] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt. [2024-11-29 11:10:07,633] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 0 [2024-11-29 11:09:54,870] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 13 [2024-11-29 11:09:54,881] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 14 [2024-11-29 11:09:54,886] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_12_mp_rank_00_optim_states.pt. [2024-11-29 11:09:54,886] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 12 [2024-11-29 11:09:54,896] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 8 [2024-11-29 11:10:07,688] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 0 [cnode7-008:0/16] 2024-11-29 11:10:07,705 (multiple_iter_factory:32) INFO: Building 0th iter-factory... [2024-11-29 11:09:54,925] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 15 [2024-11-29 11:09:54,927] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 10 [2024-11-29 11:09:54,928] [INFO] [torch_checkpoint_engine.py:29:load] [Torch] Loaded checkpoint from exp_owsm/s2t_train_05b_ds_raw_bpe50000/checkpoint_8/8/bf16_zero_pp_rank_11_mp_rank_00_optim_states.pt. [2024-11-29 11:09:54,928] [INFO] [engine.py:3076:_get_all_zero_checkpoint_state_dicts] successfully read 16 ZeRO state_dicts for rank 11 [2024-11-29 11:09:54,945] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 12 [2024-11-29 11:09:54,983] [INFO] [engine.py:3026:_load_zero_checkpoint] loading 16 zero partition checkpoints for rank 11 [cnode7-008:0/16] 2024-11-29 11:10:34,255 (s2t:444) INFO: Optional Data Names: ('text_prev', 'text_ctc', 'text_spk2', 'text_spk3', 'text_spk4') [cnode7-008:0/16] 2024-11-29 11:10:40,087 (abs_task:1807) INFO: [train] dataset: ESPnetDataset( speech: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/wav.scp/split.4", "type": "kaldi_ark"} text_prev: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.prev/split.4", "type": "text"} text_ctc: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.ctc/split.4", "type": "text"} text: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text/split.4", "type": "text"} preprocess: ) [cnode7-008:0/16] 2024-11-29 11:10:40,087 (abs_task:1808) INFO: [train] Batch sampler: UnsortedBatchSampler(N-batch=28521, batch_size=256, key_file=exp_owsm/s2t_stats_raw_bpe50000/splits8/speech_shape/split.4, [cnode7-008:0/16] 2024-11-29 11:10:40,089 (abs_task:1809) INFO: [train] mini-batch sizes summary: N-batch=28521, mean=256.0, min=256, max=257 [2024-11-29 11:11:02,906] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:10:50,129] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:03,064] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:03,234] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:03,861] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:04,096] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:04,202] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:10:51,421] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:04,523] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:10:52,205] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:10:52,270] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:10:52,678] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:10:52,781] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:05,659] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:10:53,460] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:10:53,731] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:50,257] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:38,044] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:51,199] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:53,100] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:53,126] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:53,317] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:53,485] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:53,858] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:41,688] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:42,026] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:42,784] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:43,199] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:43,712] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:43,823] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:11:44,118] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:02,857] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:24,700] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:38,805] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:40,031] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:40,710] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:41,618] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:42,229] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:42,249] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:42,617] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:31,312] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:31,760] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:32,626] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:34,119] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:34,264] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:34,419] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:34,474] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:12:52,516] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:12,138] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:27,581] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:28,838] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:29,022] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:30,891] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:31,251] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:33,872] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:21,200] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:21,709] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:22,102] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:35,055] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:24,082] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:24,430] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:41,637] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:29,576] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:13:36,647] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): cnode7-009:2307524:2315806 [5] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307524:2315806 [5] NCCL INFO Using network IB cnode7-008:3116462:3124970 [3] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116462:3124970 [3] NCCL INFO Using network IB cnode7-008:3116459:3124968 [0] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116459:3124968 [0] NCCL INFO Using network IB cnode7-008:3116463:3124969 [4] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116463:3124969 [4] NCCL INFO Using network IB /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): cnode7-008:3116465:3125250 [6] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116465:3125250 [6] NCCL INFO Using network IB cnode7-008:3116466:3125320 [7] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116466:3125320 [7] NCCL INFO Using network IB /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): cnode7-008:3116461:3125393 [2] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116461:3125393 [2] NCCL INFO Using network IB /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): cnode7-009:2307526:2316098 [7] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307526:2316098 [7] NCCL INFO Using network IB /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): cnode7-009:2307522:2316180 [3] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307522:2316180 [3] NCCL INFO Using network IB cnode7-009:2307521:2316254 [2] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307521:2316254 [2] NCCL INFO Using network IB /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): cnode7-008:3116464:3125472 [5] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116464:3125472 [5] NCCL INFO Using network IB /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): cnode7-009:2307519:2316470 [0] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307519:2316470 [0] NCCL INFO Using network IB cnode7-009:2307523:2316535 [4] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307523:2316535 [4] NCCL INFO Using network IB /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): cnode7-008:3116460:3125682 [1] NCCL INFO Using non-device net plugin version 0 cnode7-008:3116460:3125682 [1] NCCL INFO Using network IB /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): cnode7-009:2307520:2316681 [1] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307520:2316681 [1] NCCL INFO Using network IB /mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py:279: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. with autocast(False): cnode7-009:2307519:2316470 [0] NCCL INFO comm 0x5555c75d5950 rank 8 nranks 16 cudaDev 0 nvmlDev 0 busId 1b000 commId 0x873df36d41928cb7 - Init START cnode7-008:3116459:3124968 [0] NCCL INFO comm 0x5555c7598b80 rank 0 nranks 16 cudaDev 0 nvmlDev 0 busId 1b000 commId 0x873df36d41928cb7 - Init START cnode7-008:3116464:3125472 [5] NCCL INFO comm 0x5555d15a0c70 rank 5 nranks 16 cudaDev 5 nvmlDev 5 busId c3000 commId 0x873df36d41928cb7 - Init START cnode7-008:3116465:3125250 [6] NCCL INFO comm 0x5555ceddaed0 rank 6 nranks 16 cudaDev 6 nvmlDev 6 busId d1000 commId 0x873df36d41928cb7 - Init START cnode7-008:3116461:3125393 [2] NCCL INFO comm 0x5555cedec320 rank 2 nranks 16 cudaDev 2 nvmlDev 2 busId 52000 commId 0x873df36d41928cb7 - Init START cnode7-009:2307523:2316535 [4] NCCL INFO comm 0x5555d15c8d70 rank 12 nranks 16 cudaDev 4 nvmlDev 4 busId 9d000 commId 0x873df36d41928cb7 - Init START cnode7-008:3116463:3124969 [4] NCCL INFO comm 0x5555d15acd40 rank 4 nranks 16 cudaDev 4 nvmlDev 4 busId 9d000 commId 0x873df36d41928cb7 - Init START cnode7-009:2307526:2316098 [7] NCCL INFO comm 0x5555cede6ea0 rank 15 nranks 16 cudaDev 7 nvmlDev 7 busId df000 commId 0x873df36d41928cb7 - Init START cnode7-008:3116460:3125682 [1] NCCL INFO comm 0x5555cedfb9d0 rank 1 nranks 16 cudaDev 1 nvmlDev 1 busId 43000 commId 0x873df36d41928cb7 - Init START cnode7-008:3116466:3125320 [7] NCCL INFO comm 0x5555cedf7d80 rank 7 nranks 16 cudaDev 7 nvmlDev 7 busId df000 commId 0x873df36d41928cb7 - Init START cnode7-009:2307521:2316254 [2] NCCL INFO comm 0x5555d15c8680 rank 10 nranks 16 cudaDev 2 nvmlDev 2 busId 52000 commId 0x873df36d41928cb7 - Init START cnode7-009:2307520:2316681 [1] NCCL INFO comm 0x5555cee14cc0 rank 9 nranks 16 cudaDev 1 nvmlDev 1 busId 43000 commId 0x873df36d41928cb7 - Init START cnode7-009:2307522:2316180 [3] NCCL INFO comm 0x5555d15cfeb0 rank 11 nranks 16 cudaDev 3 nvmlDev 3 busId 61000 commId 0x873df36d41928cb7 - Init START cnode7-009:2307525:2317201 [6] NCCL INFO Using non-device net plugin version 0 cnode7-009:2307525:2317201 [6] NCCL INFO Using network IB cnode7-009:2307525:2317201 [6] NCCL INFO comm 0x5555d1598ab0 rank 14 nranks 16 cudaDev 6 nvmlDev 6 busId d1000 commId 0x873df36d41928cb7 - Init START cnode7-009:2307524:2315806 [5] NCCL INFO comm 0x5555cee1b4b0 rank 13 nranks 16 cudaDev 5 nvmlDev 5 busId c3000 commId 0x873df36d41928cb7 - Init START cnode7-008:3116462:3124970 [3] NCCL INFO comm 0x5555d15c0a00 rank 3 nranks 16 cudaDev 3 nvmlDev 3 busId 61000 commId 0x873df36d41928cb7 - Init START cnode7-008:3116461:3125393 [2] NCCL INFO Setting affinity for GPU 2 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-008:3116461:3125393 [2] NCCL INFO NVLS multicast support is available on dev 2 cnode7-008:3116460:3125682 [1] NCCL INFO Setting affinity for GPU 1 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-008:3116460:3125682 [1] NCCL INFO NVLS multicast support is available on dev 1 cnode7-008:3116462:3124970 [3] NCCL INFO Setting affinity for GPU 3 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-008:3116462:3124970 [3] NCCL INFO NVLS multicast support is available on dev 3 cnode7-009:2307520:2316681 [1] NCCL INFO Setting affinity for GPU 1 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-009:2307520:2316681 [1] NCCL INFO NVLS multicast support is available on dev 1 cnode7-009:2307525:2317201 [6] NCCL INFO Setting affinity for GPU 6 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-009:2307525:2317201 [6] NCCL INFO NVLS multicast support is available on dev 6 cnode7-008:3116459:3124968 [0] NCCL INFO Setting affinity for GPU 0 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-008:3116459:3124968 [0] NCCL INFO NVLS multicast support is available on dev 0 cnode7-008:3116464:3125472 [5] NCCL INFO Setting affinity for GPU 5 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-008:3116464:3125472 [5] NCCL INFO NVLS multicast support is available on dev 5 cnode7-008:3116465:3125250 [6] NCCL INFO Setting affinity for GPU 6 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-008:3116465:3125250 [6] NCCL INFO NVLS multicast support is available on dev 6 cnode7-008:3116463:3124969 [4] NCCL INFO Setting affinity for GPU 4 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-008:3116463:3124969 [4] NCCL INFO NVLS multicast support is available on dev 4 cnode7-009:2307526:2316098 [7] NCCL INFO Setting affinity for GPU 7 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-009:2307526:2316098 [7] NCCL INFO NVLS multicast support is available on dev 7 cnode7-008:3116466:3125320 [7] NCCL INFO Setting affinity for GPU 7 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-008:3116466:3125320 [7] NCCL INFO NVLS multicast support is available on dev 7 cnode7-009:2307524:2315806 [5] NCCL INFO Setting affinity for GPU 5 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-009:2307524:2315806 [5] NCCL INFO NVLS multicast support is available on dev 5 cnode7-009:2307523:2316535 [4] NCCL INFO Setting affinity for GPU 4 to ffffffff,ffffff00,00000000,0000ffff,ffffffff,ff000000,00000000 cnode7-009:2307523:2316535 [4] NCCL INFO NVLS multicast support is available on dev 4 cnode7-009:2307521:2316254 [2] NCCL INFO Setting affinity for GPU 2 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-009:2307521:2316254 [2] NCCL INFO NVLS multicast support is available on dev 2 cnode7-009:2307522:2316180 [3] NCCL INFO Setting affinity for GPU 3 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-009:2307522:2316180 [3] NCCL INFO NVLS multicast support is available on dev 3 cnode7-009:2307525:2317201 [6] NCCL INFO comm 0x5555d1598ab0 rank 14 nRanks 16 nNodes 2 localRanks 8 localRank 6 MNNVL 0 cnode7-009:2307525:2317201 [6] NCCL INFO Trees [0] 15/-1/-1->14->13 [1] 15/-1/-1->14->13 [2] 15/-1/-1->14->13 [3] 15/-1/-1->14->13 [4] 15/-1/-1->14->13 [5] 15/-1/-1->14->13 [6] 15/-1/-1->14->6 [7] -1/-1/-1->14->13 [8] 15/-1/-1->14->13 [9] 15/-1/-1->14->13 [10] 15/-1/-1->14->13 [11] 15/-1/-1->14->13 [12] 15/-1/-1->14->13 [13] 15/-1/-1->14->13 [14] 15/6/-1->14->-1 [15] -1/-1/-1->14->13 cnode7-009:2307525:2317201 [6] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116462:3124970 [3] NCCL INFO comm 0x5555d15c0a00 rank 3 nRanks 16 nNodes 2 localRanks 8 localRank 3 MNNVL 0 cnode7-008:3116462:3124970 [3] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116462:3124970 [3] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116462:3124970 [3] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116462:3124970 [3] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116462:3124970 [3] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116462:3124970 [3] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116462:3124970 [3] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116462:3124970 [3] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116462:3124970 [3] NCCL INFO Trees [0] 4/-1/-1->3->2 [1] 4/-1/-1->3->2 [2] 4/-1/-1->3->2 [3] 4/11/-1->3->-1 [4] -1/-1/-1->3->2 [5] 4/-1/-1->3->2 [6] 4/-1/-1->3->2 [7] 4/-1/-1->3->2 [8] 4/-1/-1->3->2 [9] 4/-1/-1->3->2 [10] 4/-1/-1->3->2 [11] 4/-1/-1->3->11 [12] -1/-1/-1->3->2 [13] 4/-1/-1->3->2 [14] 4/-1/-1->3->2 [15] 4/-1/-1->3->2 cnode7-008:3116462:3124970 [3] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116464:3125472 [5] NCCL INFO comm 0x5555d15a0c70 rank 5 nRanks 16 nNodes 2 localRanks 8 localRank 5 MNNVL 0 cnode7-008:3116464:3125472 [5] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116464:3125472 [5] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116464:3125472 [5] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116464:3125472 [5] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116464:3125472 [5] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116464:3125472 [5] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116464:3125472 [5] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116464:3125472 [5] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116464:3125472 [5] NCCL INFO Trees [0] 6/-1/-1->5->4 [1] 6/-1/-1->5->4 [2] 6/-1/-1->5->4 [3] 6/-1/-1->5->4 [4] 6/-1/-1->5->4 [5] 6/13/-1->5->-1 [6] -1/-1/-1->5->4 [7] 6/-1/-1->5->4 [8] 6/-1/-1->5->4 [9] 6/-1/-1->5->4 [10] 6/-1/-1->5->4 [11] 6/-1/-1->5->4 [12] 6/-1/-1->5->4 [13] 6/-1/-1->5->13 [14] -1/-1/-1->5->4 [15] 6/-1/-1->5->4 cnode7-008:3116464:3125472 [5] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116461:3125393 [2] NCCL INFO comm 0x5555cedec320 rank 2 nRanks 16 nNodes 2 localRanks 8 localRank 2 MNNVL 0 cnode7-008:3116461:3125393 [2] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116461:3125393 [2] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116461:3125393 [2] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116461:3125393 [2] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116461:3125393 [2] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116461:3125393 [2] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116461:3125393 [2] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116461:3125393 [2] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116461:3125393 [2] NCCL INFO Trees [0] 3/-1/-1->2->1 [1] 3/-1/-1->2->1 [2] 3/10/-1->2->-1 [3] -1/-1/-1->2->1 [4] 3/-1/-1->2->1 [5] 3/-1/-1->2->1 [6] 3/-1/-1->2->1 [7] 3/-1/-1->2->1 [8] 3/-1/-1->2->1 [9] 3/-1/-1->2->1 [10] 3/-1/-1->2->10 [11] -1/-1/-1->2->1 [12] 3/-1/-1->2->1 [13] 3/-1/-1->2->1 [14] 3/-1/-1->2->1 [15] 3/-1/-1->2->1 cnode7-008:3116461:3125393 [2] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307524:2315806 [5] NCCL INFO comm 0x5555cee1b4b0 rank 13 nRanks 16 nNodes 2 localRanks 8 localRank 5 MNNVL 0 cnode7-009:2307524:2315806 [5] NCCL INFO Trees [0] 14/-1/-1->13->12 [1] 14/-1/-1->13->12 [2] 14/-1/-1->13->12 [3] 14/-1/-1->13->12 [4] 14/-1/-1->13->12 [5] 14/-1/-1->13->5 [6] -1/-1/-1->13->12 [7] 14/-1/-1->13->12 [8] 14/-1/-1->13->12 [9] 14/-1/-1->13->12 [10] 14/-1/-1->13->12 [11] 14/-1/-1->13->12 [12] 14/-1/-1->13->12 [13] 14/5/-1->13->-1 [14] -1/-1/-1->13->12 [15] 14/-1/-1->13->12 cnode7-009:2307524:2315806 [5] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116465:3125250 [6] NCCL INFO comm 0x5555ceddaed0 rank 6 nRanks 16 nNodes 2 localRanks 8 localRank 6 MNNVL 0 cnode7-008:3116465:3125250 [6] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116465:3125250 [6] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116465:3125250 [6] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116465:3125250 [6] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116465:3125250 [6] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116465:3125250 [6] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116465:3125250 [6] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116465:3125250 [6] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116465:3125250 [6] NCCL INFO Trees [0] 7/-1/-1->6->5 [1] 7/-1/-1->6->5 [2] 7/-1/-1->6->5 [3] 7/-1/-1->6->5 [4] 7/-1/-1->6->5 [5] 7/-1/-1->6->5 [6] 7/14/-1->6->-1 [7] -1/-1/-1->6->5 [8] 7/-1/-1->6->5 [9] 7/-1/-1->6->5 [10] 7/-1/-1->6->5 [11] 7/-1/-1->6->5 [12] 7/-1/-1->6->5 [13] 7/-1/-1->6->5 [14] 7/-1/-1->6->14 [15] -1/-1/-1->6->5 cnode7-008:3116465:3125250 [6] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307523:2316535 [4] NCCL INFO comm 0x5555d15c8d70 rank 12 nRanks 16 nNodes 2 localRanks 8 localRank 4 MNNVL 0 cnode7-009:2307523:2316535 [4] NCCL INFO Trees [0] 13/-1/-1->12->11 [1] 13/-1/-1->12->11 [2] 13/-1/-1->12->11 [3] 13/-1/-1->12->11 [4] 13/-1/-1->12->4 [5] -1/-1/-1->12->11 [6] 13/-1/-1->12->11 [7] 13/-1/-1->12->11 [8] 13/-1/-1->12->11 [9] 13/-1/-1->12->11 [10] 13/-1/-1->12->11 [11] 13/-1/-1->12->11 [12] 13/4/-1->12->-1 [13] -1/-1/-1->12->11 [14] 13/-1/-1->12->11 [15] 13/-1/-1->12->11 cnode7-009:2307523:2316535 [4] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307526:2316098 [7] NCCL INFO comm 0x5555cede6ea0 rank 15 nRanks 16 nNodes 2 localRanks 8 localRank 7 MNNVL 0 cnode7-009:2307526:2316098 [7] NCCL INFO Trees [0] -1/-1/-1->15->14 [1] 8/-1/-1->15->14 [2] 8/-1/-1->15->14 [3] 8/-1/-1->15->14 [4] 8/-1/-1->15->14 [5] 8/-1/-1->15->14 [6] 8/-1/-1->15->14 [7] 8/-1/-1->15->7 [8] -1/-1/-1->15->14 [9] 8/-1/-1->15->14 [10] 8/-1/-1->15->14 [11] 8/-1/-1->15->14 [12] 8/-1/-1->15->14 [13] 8/-1/-1->15->14 [14] 8/-1/-1->15->14 [15] 8/7/-1->15->-1 cnode7-009:2307526:2316098 [7] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307521:2316254 [2] NCCL INFO comm 0x5555d15c8680 rank 10 nRanks 16 nNodes 2 localRanks 8 localRank 2 MNNVL 0 cnode7-009:2307521:2316254 [2] NCCL INFO Trees [0] 11/-1/-1->10->9 [1] 11/-1/-1->10->9 [2] 11/-1/-1->10->2 [3] -1/-1/-1->10->9 [4] 11/-1/-1->10->9 [5] 11/-1/-1->10->9 [6] 11/-1/-1->10->9 [7] 11/-1/-1->10->9 [8] 11/-1/-1->10->9 [9] 11/-1/-1->10->9 [10] 11/2/-1->10->-1 [11] -1/-1/-1->10->9 [12] 11/-1/-1->10->9 [13] 11/-1/-1->10->9 [14] 11/-1/-1->10->9 [15] 11/-1/-1->10->9 cnode7-009:2307521:2316254 [2] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116460:3125682 [1] NCCL INFO comm 0x5555cedfb9d0 rank 1 nRanks 16 nNodes 2 localRanks 8 localRank 1 MNNVL 0 cnode7-008:3116460:3125682 [1] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116460:3125682 [1] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116460:3125682 [1] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116460:3125682 [1] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116460:3125682 [1] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116460:3125682 [1] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116460:3125682 [1] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116460:3125682 [1] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116460:3125682 [1] NCCL INFO Trees [0] 2/-1/-1->1->0 [1] 2/9/-1->1->-1 [2] -1/-1/-1->1->0 [3] 2/-1/-1->1->0 [4] 2/-1/-1->1->0 [5] 2/-1/-1->1->0 [6] 2/-1/-1->1->0 [7] 2/-1/-1->1->0 [8] 2/-1/-1->1->0 [9] 2/-1/-1->1->9 [10] -1/-1/-1->1->0 [11] 2/-1/-1->1->0 [12] 2/-1/-1->1->0 [13] 2/-1/-1->1->0 [14] 2/-1/-1->1->0 [15] 2/-1/-1->1->0 cnode7-008:3116460:3125682 [1] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116459:3124968 [0] NCCL INFO comm 0x5555c7598b80 rank 0 nRanks 16 nNodes 2 localRanks 8 localRank 0 MNNVL 0 cnode7-008:3116459:3124968 [0] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116459:3124968 [0] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116459:3124968 [0] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116459:3124968 [0] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116459:3124968 [0] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116459:3124968 [0] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116459:3124968 [0] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116459:3124968 [0] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 00/16 : 0 7 6 5 4 3 2 1 9 10 11 12 13 14 15 8 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 01/16 : 0 8 15 14 13 12 11 10 9 1 2 3 4 5 6 7 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 02/16 : 0 7 6 5 4 3 11 12 13 14 15 8 9 10 2 1 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 03/16 : 0 1 2 10 9 8 15 14 13 12 11 3 4 5 6 7 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 04/16 : 0 7 6 5 13 14 15 8 9 10 11 12 4 3 2 1 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 05/16 : 0 1 2 3 4 12 11 10 9 8 15 14 13 5 6 7 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 06/16 : 0 7 15 8 9 10 11 12 13 14 6 5 4 3 2 1 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 07/16 : 0 1 2 3 4 5 6 14 13 12 11 10 9 8 15 7 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 08/16 : 0 7 6 5 4 3 2 1 9 10 11 12 13 14 15 8 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 09/16 : 0 8 15 14 13 12 11 10 9 1 2 3 4 5 6 7 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 10/16 : 0 7 6 5 4 3 11 12 13 14 15 8 9 10 2 1 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 11/16 : 0 1 2 10 9 8 15 14 13 12 11 3 4 5 6 7 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 12/16 : 0 7 6 5 13 14 15 8 9 10 11 12 4 3 2 1 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 13/16 : 0 1 2 3 4 12 11 10 9 8 15 14 13 5 6 7 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 14/16 : 0 7 15 8 9 10 11 12 13 14 6 5 4 3 2 1 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 15/16 : 0 1 2 3 4 5 6 14 13 12 11 10 9 8 15 7 cnode7-008:3116459:3124968 [0] NCCL INFO Trees [0] 1/8/-1->0->-1 [1] -1/-1/-1->0->7 [2] 1/-1/-1->0->7 [3] 1/-1/-1->0->7 [4] 1/-1/-1->0->7 [5] 1/-1/-1->0->7 [6] 1/-1/-1->0->7 [7] 1/-1/-1->0->7 [8] 1/-1/-1->0->8 [9] -1/-1/-1->0->7 [10] 1/-1/-1->0->7 [11] 1/-1/-1->0->7 [12] 1/-1/-1->0->7 [13] 1/-1/-1->0->7 [14] 1/-1/-1->0->7 [15] 1/-1/-1->0->7 cnode7-008:3116459:3124968 [0] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307519:2316470 [0] NCCL INFO Setting affinity for GPU 0 to ff,ffffffff,ffff0000,00000000,00ffffff,ffffffff cnode7-009:2307519:2316470 [0] NCCL INFO NVLS multicast support is available on dev 0 cnode7-009:2307519:2316470 [0] NCCL INFO comm 0x5555c75d5950 rank 8 nRanks 16 nNodes 2 localRanks 8 localRank 0 MNNVL 0 cnode7-009:2307519:2316470 [0] NCCL INFO Trees [0] 9/-1/-1->8->0 [1] -1/-1/-1->8->15 [2] 9/-1/-1->8->15 [3] 9/-1/-1->8->15 [4] 9/-1/-1->8->15 [5] 9/-1/-1->8->15 [6] 9/-1/-1->8->15 [7] 9/-1/-1->8->15 [8] 9/0/-1->8->-1 [9] -1/-1/-1->8->15 [10] 9/-1/-1->8->15 [11] 9/-1/-1->8->15 [12] 9/-1/-1->8->15 [13] 9/-1/-1->8->15 [14] 9/-1/-1->8->15 [15] 9/-1/-1->8->15 cnode7-009:2307519:2316470 [0] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307522:2316180 [3] NCCL INFO comm 0x5555d15cfeb0 rank 11 nRanks 16 nNodes 2 localRanks 8 localRank 3 MNNVL 0 cnode7-009:2307522:2316180 [3] NCCL INFO Trees [0] 12/-1/-1->11->10 [1] 12/-1/-1->11->10 [2] 12/-1/-1->11->10 [3] 12/-1/-1->11->3 [4] -1/-1/-1->11->10 [5] 12/-1/-1->11->10 [6] 12/-1/-1->11->10 [7] 12/-1/-1->11->10 [8] 12/-1/-1->11->10 [9] 12/-1/-1->11->10 [10] 12/-1/-1->11->10 [11] 12/3/-1->11->-1 [12] -1/-1/-1->11->10 [13] 12/-1/-1->11->10 [14] 12/-1/-1->11->10 [15] 12/-1/-1->11->10 cnode7-009:2307522:2316180 [3] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116463:3124969 [4] NCCL INFO comm 0x5555d15acd40 rank 4 nRanks 16 nNodes 2 localRanks 8 localRank 4 MNNVL 0 cnode7-008:3116463:3124969 [4] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116463:3124969 [4] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116463:3124969 [4] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116463:3124969 [4] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116463:3124969 [4] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116463:3124969 [4] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116463:3124969 [4] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116463:3124969 [4] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116463:3124969 [4] NCCL INFO Trees [0] 5/-1/-1->4->3 [1] 5/-1/-1->4->3 [2] 5/-1/-1->4->3 [3] 5/-1/-1->4->3 [4] 5/12/-1->4->-1 [5] -1/-1/-1->4->3 [6] 5/-1/-1->4->3 [7] 5/-1/-1->4->3 [8] 5/-1/-1->4->3 [9] 5/-1/-1->4->3 [10] 5/-1/-1->4->3 [11] 5/-1/-1->4->3 [12] 5/-1/-1->4->12 [13] -1/-1/-1->4->3 [14] 5/-1/-1->4->3 [15] 5/-1/-1->4->3 cnode7-008:3116463:3124969 [4] NCCL INFO P2P Chunksize set to 131072 cnode7-009:2307520:2316681 [1] NCCL INFO comm 0x5555cee14cc0 rank 9 nRanks 16 nNodes 2 localRanks 8 localRank 1 MNNVL 0 cnode7-009:2307520:2316681 [1] NCCL INFO Trees [0] 10/-1/-1->9->8 [1] 10/-1/-1->9->1 [2] -1/-1/-1->9->8 [3] 10/-1/-1->9->8 [4] 10/-1/-1->9->8 [5] 10/-1/-1->9->8 [6] 10/-1/-1->9->8 [7] 10/-1/-1->9->8 [8] 10/-1/-1->9->8 [9] 10/1/-1->9->-1 [10] -1/-1/-1->9->8 [11] 10/-1/-1->9->8 [12] 10/-1/-1->9->8 [13] 10/-1/-1->9->8 [14] 10/-1/-1->9->8 [15] 10/-1/-1->9->8 cnode7-009:2307520:2316681 [1] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116466:3125320 [7] NCCL INFO comm 0x5555cedf7d80 rank 7 nRanks 16 nNodes 2 localRanks 8 localRank 7 MNNVL 0 cnode7-008:3116466:3125320 [7] NCCL INFO NVLS Head 0: 0 8 cnode7-008:3116466:3125320 [7] NCCL INFO NVLS Head 1: 1 9 cnode7-008:3116466:3125320 [7] NCCL INFO NVLS Head 2: 2 10 cnode7-008:3116466:3125320 [7] NCCL INFO NVLS Head 3: 3 11 cnode7-008:3116466:3125320 [7] NCCL INFO NVLS Head 4: 4 12 cnode7-008:3116466:3125320 [7] NCCL INFO NVLS Head 5: 5 13 cnode7-008:3116466:3125320 [7] NCCL INFO NVLS Head 6: 6 14 cnode7-008:3116466:3125320 [7] NCCL INFO NVLS Head 7: 7 15 cnode7-008:3116466:3125320 [7] NCCL INFO Trees [0] -1/-1/-1->7->6 [1] 0/-1/-1->7->6 [2] 0/-1/-1->7->6 [3] 0/-1/-1->7->6 [4] 0/-1/-1->7->6 [5] 0/-1/-1->7->6 [6] 0/-1/-1->7->6 [7] 0/15/-1->7->-1 [8] -1/-1/-1->7->6 [9] 0/-1/-1->7->6 [10] 0/-1/-1->7->6 [11] 0/-1/-1->7->6 [12] 0/-1/-1->7->6 [13] 0/-1/-1->7->6 [14] 0/-1/-1->7->6 [15] 0/-1/-1->7->15 cnode7-008:3116466:3125320 [7] NCCL INFO P2P Chunksize set to 131072 cnode7-008:3116464:3125472 [5] NCCL INFO Channel 01/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 03/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 05/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 07/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 09/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 11/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 13/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 15/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 01/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 03/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 05/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 09/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 11/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 13/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 01/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 05/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 07/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 09/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 13/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 15/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 02/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 10/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 03/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 11/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 00/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 02/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 06/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 08/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 10/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 14/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 01/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 03/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 07/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 09/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 11/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 15/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 01/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 03/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 05/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 07/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 09/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 11/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 13/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 15/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 01/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 09/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 00/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 08/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 00/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 04/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 06/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 08/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 12/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 14/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 00/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 02/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 04/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 06/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 08/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 10/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 12/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 14/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 00/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 02/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 04/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 06/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 08/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 10/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 12/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 14/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 02/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 10/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 03/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 11/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 00/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 02/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 04/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 08/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 10/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 12/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 00/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 02/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 04/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 06/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 08/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 10/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 12/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 14/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 04/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 12/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 05/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 13/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 01/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 03/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 05/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 07/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 09/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 11/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 13/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 15/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 03/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 11/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 02/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 10/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 00/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 04/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 06/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 08/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 12/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 14/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 02/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 04/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 06/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 10/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 12/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 14/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 01/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 03/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 05/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 07/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 09/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 11/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 13/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 15/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 01/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 09/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 00/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 08/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 03/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 05/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 07/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 11/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 13/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 15/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 00/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 02/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 04/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 06/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 08/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 10/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 12/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 14/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 00/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 08/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 01/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 09/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 05/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 13/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 04/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 12/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 00/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 02/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 06/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 08/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 10/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 14/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 06/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 14/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 07/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 15/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 00/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 02/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 04/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 06/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 08/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 10/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 12/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 14/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 00/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 02/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 04/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 06/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 08/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 10/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 12/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 14/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 05/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 13/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 04/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 12/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 01/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 03/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 05/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 07/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 09/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 11/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 13/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 15/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 04/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 12/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 05/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 13/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 00/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 02/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 04/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 06/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 08/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 10/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 12/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 14/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 06/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 14/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 07/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 15/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 00/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 02/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 04/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 06/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 08/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 10/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 12/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 14/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 01/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 03/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 05/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 09/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 11/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 13/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 02/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 04/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 06/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 10/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 12/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 14/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 07/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 15/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 06/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 14/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 01/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 03/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 05/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 07/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 09/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 11/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 13/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 15/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 00/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 02/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 04/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 08/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 10/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 12/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 03/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 11/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 02/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 10/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 01/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 03/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 05/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 07/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 09/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 11/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 13/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 15/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 00/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 08/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 01/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 09/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 03/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 05/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 07/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 11/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 13/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 15/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 01/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 05/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 07/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 09/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 13/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 15/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 07/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 15/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 06/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 14/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 01/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 03/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 05/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 07/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 09/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 11/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 13/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 15/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 01/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 03/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 07/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 09/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 11/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 15/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Connected all rings cnode7-009:2307521:2316254 [2] NCCL INFO Channel 01/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 02/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Connected all rings cnode7-009:2307524:2315806 [5] NCCL INFO Connected all rings cnode7-009:2307524:2315806 [5] NCCL INFO Channel 01/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 03/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 05/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 07/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 09/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 11/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 13/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 15/0 : 13[5] -> 14[6] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Connected all rings cnode7-008:3116462:3124970 [3] NCCL INFO Channel 00/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 02/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 06/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 08/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 10/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 14/0 : 3[3] -> 4[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 03/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 11/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Connected all rings cnode7-009:2307519:2316470 [0] NCCL INFO Channel 00/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 03/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 05/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 07/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 08/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 11/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 13/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 15/0 : 8[0] -> 9[1] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 02/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 04/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 06/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 10/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 12/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 14/0 : 8[0] -> 15[7] via P2P/CUMEM cnode7-009:2307519:2316470 [0] NCCL INFO Channel 00/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 08/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Connected all rings cnode7-008:3116459:3124968 [0] NCCL INFO Channel 00/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 02/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 04/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 06/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 08/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 10/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 12/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 14/0 : 0[0] -> 1[1] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 01/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 03/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 05/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 07/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 09/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 11/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 13/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 15/0 : 0[0] -> 7[7] via P2P/CUMEM cnode7-008:3116459:3124968 [0] NCCL INFO Channel 00/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 08/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Connected all rings cnode7-008:3116464:3125472 [5] NCCL INFO Channel 00/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 02/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 04/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 08/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 10/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 12/0 : 5[5] -> 6[6] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 05/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 13/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 01/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 03/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 04/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 07/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 09/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 11/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 12/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116464:3125472 [5] NCCL INFO Channel 15/0 : 5[5] -> 4[4] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Connected all rings cnode7-008:3116465:3125250 [6] NCCL INFO Channel 00/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 02/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 04/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 06/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 08/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 10/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 12/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 14/0 : 6[6] -> 7[7] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 06/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 14/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 01/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 03/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 05/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 07/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 09/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 11/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 13/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116465:3125250 [6] NCCL INFO Channel 15/0 : 6[6] -> 5[5] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Connected all rings cnode7-008:3116461:3125393 [2] NCCL INFO Channel 00/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 02/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 04/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 06/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 08/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 10/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 12/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 14/0 : 2[2] -> 3[3] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 02/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 10/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 01/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 03/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 05/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 07/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 09/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 11/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 13/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-008:3116461:3125393 [2] NCCL INFO Channel 15/0 : 2[2] -> 1[1] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Connected all rings cnode7-009:2307523:2316535 [4] NCCL INFO Channel 01/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 03/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 04/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 07/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 09/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 11/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 12/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 15/0 : 12[4] -> 13[5] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 04/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 12/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 00/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 02/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 06/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 08/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 10/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-009:2307523:2316535 [4] NCCL INFO Channel 14/0 : 12[4] -> 11[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Connected all rings cnode7-008:3116463:3124969 [4] NCCL INFO Channel 00/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 02/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 04/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 06/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 08/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 10/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 12/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 14/0 : 4[4] -> 5[5] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 04/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 12/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 01/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 03/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 05/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 07/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 09/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 11/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 13/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-008:3116463:3124969 [4] NCCL INFO Channel 15/0 : 4[4] -> 3[3] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Connected all rings cnode7-009:2307526:2316098 [7] NCCL INFO Channel 07/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 15/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 01/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 03/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 05/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 07/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 09/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 11/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 13/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 15/0 : 15[7] -> 8[0] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 00/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 02/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 04/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 06/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 08/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 10/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 12/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-009:2307526:2316098 [7] NCCL INFO Channel 14/0 : 15[7] -> 14[6] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Connected all rings cnode7-008:3116460:3125682 [1] NCCL INFO Channel 00/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 04/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 06/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 08/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 12/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 14/0 : 1[1] -> 2[2] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 01/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 09/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 00/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 03/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 05/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 07/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 08/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 11/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 13/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116460:3125682 [1] NCCL INFO Channel 15/0 : 1[1] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Connected all rings cnode7-008:3116466:3125320 [7] NCCL INFO Channel 07/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 15/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 02/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 04/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 06/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 10/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 12/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 14/0 : 7[7] -> 0[0] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 01/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 03/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 05/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 06/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 09/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 11/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 13/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-008:3116466:3125320 [7] NCCL INFO Channel 14/0 : 7[7] -> 6[6] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 05/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 07/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 09/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 10/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 13/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 15/0 : 10[2] -> 11[3] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 02/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 10/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 00/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 04/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 06/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 08/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 12/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Channel 14/0 : 10[2] -> 9[1] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Connected all rings cnode7-009:2307520:2316681 [1] NCCL INFO Channel 01/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 03/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 05/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 07/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 09/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 11/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 13/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 15/0 : 9[1] -> 10[2] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 01/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 09/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 00/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 02/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 04/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 06/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 08/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 10/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 12/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307520:2316681 [1] NCCL INFO Channel 14/0 : 9[1] -> 8[0] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 01/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 03/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 05/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 07/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 09/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 11/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 13/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 15/0 : 11[3] -> 12[4] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 03/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 11/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 00/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 02/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 04/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 06/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 08/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 10/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 12/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307522:2316180 [3] NCCL INFO Channel 14/0 : 11[3] -> 10[2] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Connected all rings cnode7-009:2307525:2317201 [6] NCCL INFO Channel 01/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 03/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 05/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 06/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 09/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 11/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 13/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 14/0 : 14[6] -> 15[7] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 06/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 14/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 00/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 02/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 04/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 08/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 10/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307525:2317201 [6] NCCL INFO Channel 12/0 : 14[6] -> 13[5] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 05/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 13/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 00/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 02/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 04/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 06/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 08/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 10/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 12/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-009:2307524:2315806 [5] NCCL INFO Channel 14/0 : 13[5] -> 12[4] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 01/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 02/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 05/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 07/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 09/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 10/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 13/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-008:3116462:3124970 [3] NCCL INFO Channel 15/0 : 3[3] -> 2[2] via P2P/CUMEM cnode7-009:2307521:2316254 [2] NCCL INFO Connected all trees cnode7-009:2307522:2316180 [3] NCCL INFO Connected all trees cnode7-008:3116462:3124970 [3] NCCL INFO Connected all trees cnode7-009:2307521:2316254 [2] NCCL INFO NVLS comm 0x5555d15c8680 headRank 2 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307522:2316180 [3] NCCL INFO NVLS comm 0x5555d15cfeb0 headRank 3 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116461:3125393 [2] NCCL INFO Connected all trees cnode7-008:3116464:3125472 [5] NCCL INFO Connected all trees cnode7-008:3116463:3124969 [4] NCCL INFO Connected all trees cnode7-009:2307524:2315806 [5] NCCL INFO Connected all trees cnode7-009:2307523:2316535 [4] NCCL INFO Connected all trees cnode7-009:2307525:2317201 [6] NCCL INFO Connected all trees cnode7-009:2307526:2316098 [7] NCCL INFO Connected all trees cnode7-009:2307519:2316470 [0] NCCL INFO Connected all trees cnode7-009:2307520:2316681 [1] NCCL INFO Connected all trees cnode7-008:3116465:3125250 [6] NCCL INFO Connected all trees cnode7-008:3116466:3125320 [7] NCCL INFO Connected all trees cnode7-008:3116459:3124968 [0] NCCL INFO Connected all trees cnode7-008:3116460:3125682 [1] NCCL INFO Connected all trees cnode7-008:3116462:3124970 [3] NCCL INFO NVLS comm 0x5555d15c0a00 headRank 3 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307519:2316470 [0] NCCL INFO NVLS comm 0x5555c75d5950 headRank 0 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116464:3125472 [5] NCCL INFO NVLS comm 0x5555d15a0c70 headRank 5 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116465:3125250 [6] NCCL INFO NVLS comm 0x5555ceddaed0 headRank 6 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116461:3125393 [2] NCCL INFO NVLS comm 0x5555cedec320 headRank 2 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307523:2316535 [4] NCCL INFO NVLS comm 0x5555d15c8d70 headRank 4 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116463:3124969 [4] NCCL INFO NVLS comm 0x5555d15acd40 headRank 4 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307526:2316098 [7] NCCL INFO NVLS comm 0x5555cede6ea0 headRank 7 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116460:3125682 [1] NCCL INFO NVLS comm 0x5555cedfb9d0 headRank 1 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116466:3125320 [7] NCCL INFO NVLS comm 0x5555cedf7d80 headRank 7 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307520:2316681 [1] NCCL INFO NVLS comm 0x5555cee14cc0 headRank 1 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307525:2317201 [6] NCCL INFO NVLS comm 0x5555d1598ab0 headRank 6 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307524:2315806 [5] NCCL INFO NVLS comm 0x5555cee1b4b0 headRank 5 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-009:2307524:2315806 [5] NCCL INFO Channel 00/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 01/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 02/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 03/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 06/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 07/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 08/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 09/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 10/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 11/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 14/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 15/0 : 5[5] -> 13[5] [receive] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 00/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 01/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 02/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 03/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 04/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 06/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 07/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 08/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 09/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 10/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 11/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 12/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 14/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307524:2315806 [5] NCCL INFO Channel 15/0 : 13[5] -> 5[5] [send] via NET/IB/6/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 02/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 03/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 04/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 05/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 06/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 07/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 10/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 11/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 12/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 13/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 14/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 15/0 : 0[0] -> 8[0] [receive] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 01/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 02/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 03/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 04/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 05/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 06/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 07/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 09/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 10/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 11/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 12/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 13/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 14/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-009:2307519:2316470 [0] NCCL INFO Channel 15/0 : 8[0] -> 0[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO NVLS comm 0x5555c7598b80 headRank 0 nHeads 8 buffSize 4194304 memSize 2097152 nvlsPerRankSize 201326592 nvlsTotalSize 1610612736 cnode7-008:3116459:3124968 [0] NCCL INFO Channel 01/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 02/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 03/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 04/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 05/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 06/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 07/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 09/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 10/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 11/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 12/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 13/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 14/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 15/0 : 8[0] -> 0[0] [receive] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 02/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 03/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 04/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 05/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 06/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 07/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 10/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 11/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 12/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 13/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 14/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116459:3124968 [0] NCCL INFO Channel 15/0 : 0[0] -> 8[0] [send] via NET/IB/0/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 00/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 01/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 02/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 03/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 04/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 06/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 07/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 08/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 09/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 10/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 11/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 12/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 14/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 15/0 : 13[5] -> 5[5] [receive] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 00/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 01/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 02/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 03/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 06/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 07/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 08/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 09/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 10/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 11/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 14/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116464:3125472 [5] NCCL INFO Channel 15/0 : 5[5] -> 13[5] [send] via NET/IB/6/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 00/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 01/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 02/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 03/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 04/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 05/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 07/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 08/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 09/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 10/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 11/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 12/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 13/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 15/0 : 14[6] -> 6[6] [receive] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 00/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 01/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 02/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 03/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 04/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 05/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 08/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 09/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 10/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 11/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 12/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116465:3125250 [6] NCCL INFO Channel 13/0 : 6[6] -> 14[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 00/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 01/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 03/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 04/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 05/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 06/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 07/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 08/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 09/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 11/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 12/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 13/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 14/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 15/0 : 10[2] -> 2[2] [receive] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 00/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 01/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 04/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 05/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 06/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 07/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 08/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 09/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 12/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 13/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 14/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-008:3116461:3125393 [2] NCCL INFO Channel 15/0 : 2[2] -> 10[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 00/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 01/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 02/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 03/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 06/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 07/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 08/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 09/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 10/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 11/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 14/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 15/0 : 4[4] -> 12[4] [receive] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 00/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 01/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 02/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 03/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 05/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 06/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 07/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 08/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 09/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 10/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 11/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 13/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 14/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Channel 15/0 : 12[4] -> 4[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 00/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 01/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 02/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 03/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 05/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 06/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 07/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 08/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 09/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 10/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 11/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 13/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 14/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 15/0 : 12[4] -> 4[4] [receive] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 00/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 01/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 02/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 03/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 06/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 07/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 08/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 09/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 10/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 11/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 14/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-008:3116463:3124969 [4] NCCL INFO Channel 15/0 : 4[4] -> 12[4] [send] via NET/IB/4/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 00/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 01/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 02/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 03/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 04/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 05/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 08/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 09/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 10/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 11/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 12/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 13/0 : 7[7] -> 15[7] [receive] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 00/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 01/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 02/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 03/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 04/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 05/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 06/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 08/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 09/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 10/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 11/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 12/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 13/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307526:2316098 [7] NCCL INFO Channel 14/0 : 15[7] -> 7[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 00/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 02/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 03/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 04/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 05/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 06/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 07/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 08/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 10/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 11/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 12/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 13/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 14/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 15/0 : 9[1] -> 1[1] [receive] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 02/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 03/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 04/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 05/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 06/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 07/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 10/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 11/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 12/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 13/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 14/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116460:3125682 [1] NCCL INFO Channel 15/0 : 1[1] -> 9[1] [send] via NET/IB/1/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 00/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 01/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 02/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 03/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 04/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 05/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 06/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 08/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 09/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 10/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 11/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 12/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 13/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 14/0 : 15[7] -> 7[7] [receive] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 00/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 01/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 02/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 03/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 04/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 05/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 08/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 09/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 10/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 11/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 12/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-008:3116466:3125320 [7] NCCL INFO Channel 13/0 : 7[7] -> 15[7] [send] via NET/IB/8/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 00/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 01/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 04/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 05/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 06/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 07/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 08/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 09/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 12/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 13/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 14/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 15/0 : 2[2] -> 10[2] [receive] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 00/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 01/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 03/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 04/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 05/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 06/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 07/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 08/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 09/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 11/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 12/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 13/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 14/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307521:2316254 [2] NCCL INFO Channel 15/0 : 10[2] -> 2[2] [send] via NET/IB/2/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 02/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 03/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 04/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 05/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 06/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 07/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 10/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 11/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 12/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 13/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 14/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 15/0 : 1[1] -> 9[1] [receive] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 00/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 02/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 03/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 04/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 05/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 06/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 07/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 08/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 10/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 11/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 12/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 13/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 14/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307520:2316681 [1] NCCL INFO Channel 15/0 : 9[1] -> 1[1] [send] via NET/IB/1/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 00/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 01/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 04/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 05/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 06/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 07/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 08/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 09/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 12/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 13/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 14/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 15/0 : 3[3] -> 11[3] [receive] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 00/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 01/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 02/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 04/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 05/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 06/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 07/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 08/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 09/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 10/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 12/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 13/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 14/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307522:2316180 [3] NCCL INFO Channel 15/0 : 11[3] -> 3[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 00/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 01/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 02/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 03/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 04/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 05/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 08/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 09/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 10/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 11/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 12/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 13/0 : 6[6] -> 14[6] [receive] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 00/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 01/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 02/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 03/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 04/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 05/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 07/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 08/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 09/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 10/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 11/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 12/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 13/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-009:2307525:2317201 [6] NCCL INFO Channel 15/0 : 14[6] -> 6[6] [send] via NET/IB/7/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 00/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 01/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 02/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 04/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 05/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 06/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 07/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 08/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 09/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 10/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 12/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 13/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 14/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 15/0 : 11[3] -> 3[3] [receive] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 00/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 01/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 04/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 05/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 06/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 07/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 08/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 09/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 12/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 13/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 14/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-008:3116462:3124970 [3] NCCL INFO Channel 15/0 : 3[3] -> 11[3] [send] via NET/IB/3/GDRDMA cnode7-009:2307523:2316535 [4] NCCL INFO Connected NVLS tree cnode7-009:2307523:2316535 [4] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307523:2316535 [4] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116463:3124969 [4] NCCL INFO Connected NVLS tree cnode7-008:3116463:3124969 [4] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116463:3124969 [4] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116460:3125682 [1] NCCL INFO Connected NVLS tree cnode7-008:3116460:3125682 [1] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116460:3125682 [1] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307520:2316681 [1] NCCL INFO Connected NVLS tree cnode7-009:2307520:2316681 [1] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307520:2316681 [1] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307522:2316180 [3] NCCL INFO Connected NVLS tree cnode7-009:2307522:2316180 [3] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307522:2316180 [3] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307525:2317201 [6] NCCL INFO Connected NVLS tree cnode7-009:2307525:2317201 [6] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307525:2317201 [6] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116459:3124968 [0] NCCL INFO Connected NVLS tree cnode7-008:3116459:3124968 [0] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116459:3124968 [0] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116462:3124970 [3] NCCL INFO Connected NVLS tree cnode7-008:3116462:3124970 [3] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116462:3124970 [3] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307524:2315806 [5] NCCL INFO Connected NVLS tree cnode7-009:2307524:2315806 [5] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307524:2315806 [5] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307519:2316470 [0] NCCL INFO Connected NVLS tree cnode7-009:2307519:2316470 [0] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307519:2316470 [0] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116465:3125250 [6] NCCL INFO Connected NVLS tree cnode7-008:3116465:3125250 [6] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116465:3125250 [6] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307521:2316254 [2] NCCL INFO Connected NVLS tree cnode7-009:2307521:2316254 [2] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307521:2316254 [2] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-009:2307526:2316098 [7] NCCL INFO Connected NVLS tree cnode7-009:2307526:2316098 [7] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-009:2307526:2316098 [7] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116461:3125393 [2] NCCL INFO Connected NVLS tree cnode7-008:3116461:3125393 [2] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116461:3125393 [2] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116466:3125320 [7] NCCL INFO Connected NVLS tree cnode7-008:3116466:3125320 [7] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116466:3125320 [7] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116464:3125472 [5] NCCL INFO Connected NVLS tree cnode7-008:3116464:3125472 [5] NCCL INFO threadThresholds 8/8/64 | 128/8/64 | 512 | 512 cnode7-008:3116464:3125472 [5] NCCL INFO 16 coll channels, 0 collnet channels, 16 nvls channels, 16 p2p channels, 2 p2p channels per peer cnode7-008:3116459:3124968 [0] NCCL INFO comm 0x5555c7598b80 rank 0 nranks 16 cudaDev 0 nvmlDev 0 busId 1b000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-008:3116464:3125472 [5] NCCL INFO comm 0x5555d15a0c70 rank 5 nranks 16 cudaDev 5 nvmlDev 5 busId c3000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-009:2307524:2315806 [5] NCCL INFO comm 0x5555cee1b4b0 rank 13 nranks 16 cudaDev 5 nvmlDev 5 busId c3000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-009:2307521:2316254 [2] NCCL INFO comm 0x5555d15c8680 rank 10 nranks 16 cudaDev 2 nvmlDev 2 busId 52000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-009:2307526:2316098 [7] NCCL INFO comm 0x5555cede6ea0 rank 15 nranks 16 cudaDev 7 nvmlDev 7 busId df000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-009:2307525:2317201 [6] NCCL INFO comm 0x5555d1598ab0 rank 14 nranks 16 cudaDev 6 nvmlDev 6 busId d1000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-009:2307520:2316681 [1] NCCL INFO comm 0x5555cee14cc0 rank 9 nranks 16 cudaDev 1 nvmlDev 1 busId 43000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-009:2307519:2316470 [0] NCCL INFO comm 0x5555c75d5950 rank 8 nranks 16 cudaDev 0 nvmlDev 0 busId 1b000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-009:2307522:2316180 [3] NCCL INFO comm 0x5555d15cfeb0 rank 11 nranks 16 cudaDev 3 nvmlDev 3 busId 61000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-009:2307523:2316535 [4] NCCL INFO comm 0x5555d15c8d70 rank 12 nranks 16 cudaDev 4 nvmlDev 4 busId 9d000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-008:3116465:3125250 [6] NCCL INFO comm 0x5555ceddaed0 rank 6 nranks 16 cudaDev 6 nvmlDev 6 busId d1000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-008:3116460:3125682 [1] NCCL INFO comm 0x5555cedfb9d0 rank 1 nranks 16 cudaDev 1 nvmlDev 1 busId 43000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-008:3116461:3125393 [2] NCCL INFO comm 0x5555cedec320 rank 2 nranks 16 cudaDev 2 nvmlDev 2 busId 52000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-008:3116462:3124970 [3] NCCL INFO comm 0x5555d15c0a00 rank 3 nranks 16 cudaDev 3 nvmlDev 3 busId 61000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-008:3116463:3124969 [4] NCCL INFO comm 0x5555d15acd40 rank 4 nranks 16 cudaDev 4 nvmlDev 4 busId 9d000 commId 0x873df36d41928cb7 - Init COMPLETE cnode7-008:3116466:3125320 [7] NCCL INFO comm 0x5555cedf7d80 rank 7 nranks 16 cudaDev 7 nvmlDev 7 busId df000 commId 0x873df36d41928cb7 - Init COMPLETE [cnode7-008:0/16] 2024-11-29 11:15:02,620 (deepspeed_trainer:228) INFO: 9epoch:train:1-100batch: iter_time=2.293, loss_ctc=89.186, loss_att=65.424, acc=0.631, loss=72.541, grad_norm=3.894, loss_scale=1.000, learning_rate=1.414e-04, step_time=0.659 [cnode7-008:0/16] 2024-11-29 11:15:39,711 (deepspeed_trainer:228) INFO: 9epoch:train:101-200batch: iter_time=1.233e-04, loss_ctc=85.198, loss_att=67.027, acc=0.624, loss=72.480, grad_norm=3.820, loss_scale=1.000, learning_rate=1.413e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 11:16:16,645 (deepspeed_trainer:228) INFO: 9epoch:train:201-300batch: iter_time=1.318e-04, loss_ctc=89.925, loss_att=66.960, acc=0.632, loss=73.821, grad_norm=3.811, loss_scale=1.000, learning_rate=1.413e-04, step_time=0.369 [cnode7-008:0/16] 2024-11-29 11:16:54,025 (deepspeed_trainer:228) INFO: 9epoch:train:301-400batch: iter_time=1.276e-04, loss_ctc=84.805, loss_att=70.490, acc=0.636, loss=74.808, grad_norm=3.488, loss_scale=1.000, learning_rate=1.412e-04, step_time=0.373 [cnode7-008:0/16] 2024-11-29 11:17:31,004 (deepspeed_trainer:228) INFO: 9epoch:train:401-500batch: iter_time=1.254e-04, loss_ctc=85.463, loss_att=77.223, acc=0.602, loss=79.675, grad_norm=3.715, loss_scale=1.000, learning_rate=1.412e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 11:18:07,569 (deepspeed_trainer:228) INFO: 9epoch:train:501-600batch: iter_time=1.259e-04, loss_ctc=90.154, loss_att=68.318, acc=0.618, loss=74.897, grad_norm=3.842, loss_scale=1.000, learning_rate=1.411e-04, step_time=0.365 [cnode7-008:0/16] 2024-11-29 11:18:44,494 (deepspeed_trainer:228) INFO: 9epoch:train:601-700batch: iter_time=1.407e-04, loss_ctc=93.223, loss_att=72.363, acc=0.645, loss=78.593, grad_norm=4.149, loss_scale=1.000, learning_rate=1.410e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 11:19:21,496 (deepspeed_trainer:228) INFO: 9epoch:train:701-800batch: iter_time=1.311e-04, loss_ctc=106.302, loss_att=75.047, acc=0.636, loss=84.442, grad_norm=4.647, loss_scale=1.000, learning_rate=1.410e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 11:19:59,127 (deepspeed_trainer:228) INFO: 9epoch:train:801-900batch: iter_time=1.202e-04, loss_ctc=94.119, loss_att=72.131, acc=0.631, loss=78.715, grad_norm=3.623, loss_scale=1.000, learning_rate=1.409e-04, step_time=0.376 [2024-11-29 11:20:39,424] [INFO] [logging.py:129:log_dist] [Rank 0] step=121000, skipped=0, lr=[np.float64(0.00014083517608157094)], mom=[[0.9, 0.98]] [2024-11-29 11:20:39,425] [INFO] [timer.py:264:stop] epoch=0/micro_step=1000/global_step=1000, RunningAvgSamplesPerSec=43.71330280877922, CurrSamplesPerSec=46.250805500902835, MemAllocated=1.77GB, MaxMemAllocated=18.32GB [cnode7-008:0/16] 2024-11-29 11:20:39,426 (deepspeed_trainer:228) INFO: 9epoch:train:901-1000batch: iter_time=1.182e-04, loss_ctc=84.424, loss_att=66.807, acc=0.648, loss=72.101, grad_norm=3.443, loss_scale=1.000, learning_rate=1.409e-04, step_time=0.403 [cnode7-008:0/16] 2024-11-29 11:21:18,264 (deepspeed_trainer:228) INFO: 9epoch:train:1001-1100batch: iter_time=1.161e-04, loss_ctc=85.445, loss_att=71.944, acc=0.627, loss=75.987, grad_norm=3.477, loss_scale=1.000, learning_rate=1.408e-04, step_time=0.388 [cnode7-008:0/16] 2024-11-29 11:21:58,733 (deepspeed_trainer:228) INFO: 9epoch:train:1101-1200batch: iter_time=1.104e-04, loss_ctc=83.987, loss_att=69.811, acc=0.638, loss=74.095, grad_norm=3.410, loss_scale=1.000, learning_rate=1.407e-04, step_time=0.404 [cnode7-008:0/16] 2024-11-29 11:22:35,761 (deepspeed_trainer:228) INFO: 9epoch:train:1201-1300batch: iter_time=1.086e-04, loss_ctc=91.506, loss_att=73.752, acc=0.617, loss=79.102, grad_norm=3.669, loss_scale=1.000, learning_rate=1.407e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 11:23:12,785 (deepspeed_trainer:228) INFO: 9epoch:train:1301-1400batch: iter_time=1.106e-04, loss_ctc=95.406, loss_att=76.323, acc=0.626, loss=82.041, grad_norm=3.794, loss_scale=1.000, learning_rate=1.406e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 11:23:50,081 (deepspeed_trainer:228) INFO: 9epoch:train:1401-1500batch: iter_time=1.073e-04, loss_ctc=92.456, loss_att=72.656, acc=0.636, loss=78.569, grad_norm=4.010, loss_scale=1.000, learning_rate=1.406e-04, step_time=0.372 [cnode7-008:0/16] 2024-11-29 11:24:27,359 (deepspeed_trainer:228) INFO: 9epoch:train:1501-1600batch: iter_time=1.077e-04, loss_ctc=98.089, loss_att=80.267, acc=0.633, loss=85.611, grad_norm=3.717, loss_scale=1.000, learning_rate=1.405e-04, step_time=0.373 [cnode7-008:0/16] 2024-11-29 11:25:04,373 (deepspeed_trainer:228) INFO: 9epoch:train:1601-1700batch: iter_time=1.092e-04, loss_ctc=87.895, loss_att=69.287, acc=0.649, loss=74.839, grad_norm=3.757, loss_scale=1.000, learning_rate=1.405e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 11:25:41,474 (deepspeed_trainer:228) INFO: 9epoch:train:1701-1800batch: iter_time=1.089e-04, loss_ctc=89.633, loss_att=73.924, acc=0.626, loss=78.639, grad_norm=3.981, loss_scale=1.000, learning_rate=1.404e-04, step_time=0.371 [cnode7-008:0/16] 2024-11-29 11:26:13,386 (multiple_iter_factory:32) INFO: Building 1th iter-factory... [cnode7-008:0/16] 2024-11-29 11:26:39,312 (s2t:444) INFO: Optional Data Names: ('text_prev', 'text_ctc', 'text_spk2', 'text_spk3', 'text_spk4') [cnode7-008:0/16] 2024-11-29 11:26:44,765 (abs_task:1807) INFO: [train] dataset: ESPnetDataset( speech: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/wav.scp/split.5", "type": "kaldi_ark"} text_prev: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.prev/split.5", "type": "text"} text_ctc: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.ctc/split.5", "type": "text"} text: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text/split.5", "type": "text"} preprocess: ) [cnode7-008:0/16] 2024-11-29 11:26:44,765 (abs_task:1808) INFO: [train] Batch sampler: UnsortedBatchSampler(N-batch=28521, batch_size=256, key_file=exp_owsm/s2t_stats_raw_bpe50000/splits8/speech_shape/split.5, [cnode7-008:0/16] 2024-11-29 11:26:44,768 (abs_task:1809) INFO: [train] mini-batch sizes summary: N-batch=28521, mean=256.0, min=256, max=257 [2024-11-29 11:27:11,776] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:11,782] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:11,905] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:12,259] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:12,577] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:12,793] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:26:59,754] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:26:59,800] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:13,230] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:00,335] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:00,390] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:00,743] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:00,780] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:01,651] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:16,349] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:05,828] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:00,580] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:01,616] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:01,673] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:48,571] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:01,806] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:01,837] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:49,356] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:49,777] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:50,055] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:50,615] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:50,654] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:03,849] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:04,297] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:04,489] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:51,441] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:27:55,537] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:48,614] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:49,658] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:37,321] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:37,592] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:51,052] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:51,108] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:51,406] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:38,925] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:39,047] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:52,280] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:39,939] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:40,468] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:41,025] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:55,593] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:56,278] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:28:43,891] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:36,947] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:37,738] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:38,528] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:26,332] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:39,924] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:26,820] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:40,329] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:28,086] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:41,620] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:28,537] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:28,821] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:29,394] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:30,862] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:32,819] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:46,424] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:29:47,288] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [cnode7-008:0/16] 2024-11-29 11:30:23,262 (deepspeed_trainer:228) INFO: 9epoch:train:1801-1900batch: iter_time=2.431, loss_ctc=85.459, loss_att=68.531, acc=0.643, loss=73.625, grad_norm=3.461, loss_scale=1.000, learning_rate=1.403e-04, step_time=0.386 [2024-11-29 11:31:02,171] [INFO] [logging.py:129:log_dist] [Rank 0] step=122000, skipped=0, lr=[np.float64(0.00014025679984188308)], mom=[[0.9, 0.98]] [2024-11-29 11:31:02,171] [INFO] [timer.py:264:stop] epoch=0/micro_step=2000/global_step=2000, RunningAvgSamplesPerSec=43.36687941935861, CurrSamplesPerSec=46.24420818763532, MemAllocated=1.77GB, MaxMemAllocated=18.32GB [cnode7-008:0/16] 2024-11-29 11:31:02,173 (deepspeed_trainer:228) INFO: 9epoch:train:1901-2000batch: iter_time=1.044e-04, loss_ctc=86.727, loss_att=64.528, acc=0.634, loss=71.210, grad_norm=3.672, loss_scale=1.000, learning_rate=1.403e-04, step_time=0.389 [cnode7-008:0/16] 2024-11-29 11:31:40,477 (deepspeed_trainer:228) INFO: 9epoch:train:2001-2100batch: iter_time=1.065e-04, loss_ctc=83.726, loss_att=66.564, acc=0.629, loss=71.726, grad_norm=3.574, loss_scale=1.000, learning_rate=1.402e-04, step_time=0.383 [cnode7-008:0/16] 2024-11-29 11:32:19,028 (deepspeed_trainer:228) INFO: 9epoch:train:2101-2200batch: iter_time=1.082e-04, loss_ctc=88.085, loss_att=67.880, acc=0.620, loss=73.949, grad_norm=3.983, loss_scale=1.000, learning_rate=1.402e-04, step_time=0.385 [cnode7-008:0/16] 2024-11-29 11:32:55,866 (deepspeed_trainer:228) INFO: 9epoch:train:2201-2300batch: iter_time=1.085e-04, loss_ctc=86.013, loss_att=72.434, acc=0.636, loss=76.538, grad_norm=3.556, loss_scale=1.000, learning_rate=1.401e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 11:33:32,424 (deepspeed_trainer:228) INFO: 9epoch:train:2301-2400batch: iter_time=1.054e-04, loss_ctc=82.999, loss_att=71.464, acc=0.609, loss=74.905, grad_norm=3.700, loss_scale=1.000, learning_rate=1.401e-04, step_time=0.366 [cnode7-008:0/16] 2024-11-29 11:34:09,400 (deepspeed_trainer:228) INFO: 9epoch:train:2401-2500batch: iter_time=1.070e-04, loss_ctc=89.547, loss_att=67.267, acc=0.626, loss=73.910, grad_norm=3.968, loss_scale=1.000, learning_rate=1.400e-04, step_time=0.369 [cnode7-008:0/16] 2024-11-29 11:34:46,497 (deepspeed_trainer:228) INFO: 9epoch:train:2501-2600batch: iter_time=1.092e-04, loss_ctc=92.184, loss_att=71.394, acc=0.634, loss=77.628, grad_norm=3.813, loss_scale=1.000, learning_rate=1.399e-04, step_time=0.371 [cnode7-008:0/16] 2024-11-29 11:35:23,908 (deepspeed_trainer:228) INFO: 9epoch:train:2601-2700batch: iter_time=1.146e-04, loss_ctc=110.728, loss_att=75.344, acc=0.643, loss=86.000, grad_norm=5.014, loss_scale=1.000, learning_rate=1.399e-04, step_time=0.374 [cnode7-008:0/16] 2024-11-29 11:36:00,922 (deepspeed_trainer:228) INFO: 9epoch:train:2701-2800batch: iter_time=1.067e-04, loss_ctc=87.094, loss_att=64.479, acc=0.632, loss=71.285, grad_norm=3.564, loss_scale=1.000, learning_rate=1.398e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 11:36:37,554 (deepspeed_trainer:228) INFO: 9epoch:train:2801-2900batch: iter_time=1.058e-04, loss_ctc=80.828, loss_att=68.653, acc=0.640, loss=72.316, grad_norm=3.315, loss_scale=1.000, learning_rate=1.398e-04, step_time=0.366 [2024-11-29 11:37:14,356] [INFO] [logging.py:129:log_dist] [Rank 0] step=123000, skipped=0, lr=[np.float64(0.00013968549132786922)], mom=[[0.9, 0.98]] [2024-11-29 11:37:14,357] [INFO] [timer.py:264:stop] epoch=0/micro_step=3000/global_step=3000, RunningAvgSamplesPerSec=43.56885560844211, CurrSamplesPerSec=43.3351773084087, MemAllocated=1.77GB, MaxMemAllocated=18.32GB [cnode7-008:0/16] 2024-11-29 11:37:14,358 (deepspeed_trainer:228) INFO: 9epoch:train:2901-3000batch: iter_time=1.069e-04, loss_ctc=85.588, loss_att=72.759, acc=0.628, loss=76.569, grad_norm=3.607, loss_scale=1.000, learning_rate=1.397e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 11:37:50,889 (deepspeed_trainer:228) INFO: 9epoch:train:3001-3100batch: iter_time=1.052e-04, loss_ctc=89.304, loss_att=70.323, acc=0.626, loss=76.034, grad_norm=3.806, loss_scale=1.000, learning_rate=1.397e-04, step_time=0.365 [cnode7-008:0/16] 2024-11-29 11:38:27,527 (deepspeed_trainer:228) INFO: 9epoch:train:3101-3200batch: iter_time=1.088e-04, loss_ctc=90.281, loss_att=71.845, acc=0.628, loss=77.380, grad_norm=3.561, loss_scale=1.000, learning_rate=1.396e-04, step_time=0.366 [cnode7-008:0/16] 2024-11-29 11:39:04,280 (deepspeed_trainer:228) INFO: 9epoch:train:3201-3300batch: iter_time=1.097e-04, loss_ctc=88.818, loss_att=75.084, acc=0.621, loss=79.199, grad_norm=3.781, loss_scale=1.000, learning_rate=1.395e-04, step_time=0.367 [cnode7-008:0/16] 2024-11-29 11:39:41,139 (deepspeed_trainer:228) INFO: 9epoch:train:3301-3400batch: iter_time=1.060e-04, loss_ctc=95.813, loss_att=73.014, acc=0.640, loss=79.814, grad_norm=4.266, loss_scale=1.000, learning_rate=1.395e-04, step_time=0.369 [cnode7-008:0/16] 2024-11-29 11:40:19,215 (deepspeed_trainer:228) INFO: 9epoch:train:3401-3500batch: iter_time=1.053e-04, loss_ctc=90.971, loss_att=71.946, acc=0.638, loss=77.631, grad_norm=3.878, loss_scale=1.000, learning_rate=1.394e-04, step_time=0.376 [cnode7-008:0/16] 2024-11-29 11:40:57,354 (deepspeed_trainer:228) INFO: 9epoch:train:3501-3600batch: iter_time=1.031e-04, loss_ctc=90.152, loss_att=69.779, acc=0.639, loss=75.879, grad_norm=3.923, loss_scale=1.000, learning_rate=1.394e-04, step_time=0.385 [cnode7-008:0/16] 2024-11-29 11:41:37,056 (deepspeed_trainer:228) INFO: 9epoch:train:3601-3700batch: iter_time=1.082e-04, loss_ctc=88.903, loss_att=74.501, acc=0.627, loss=78.803, grad_norm=3.718, loss_scale=1.000, learning_rate=1.393e-04, step_time=0.397 [cnode7-008:0/16] 2024-11-29 11:42:00,086 (multiple_iter_factory:32) INFO: Building 2th iter-factory... [cnode7-008:0/16] 2024-11-29 11:42:26,539 (s2t:444) INFO: Optional Data Names: ('text_prev', 'text_ctc', 'text_spk2', 'text_spk3', 'text_spk4') [cnode7-008:0/16] 2024-11-29 11:42:32,059 (abs_task:1807) INFO: [train] dataset: ESPnetDataset( speech: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/wav.scp/split.6", "type": "kaldi_ark"} text_prev: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.prev/split.6", "type": "text"} text_ctc: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.ctc/split.6", "type": "text"} text: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text/split.6", "type": "text"} preprocess: ) [cnode7-008:0/16] 2024-11-29 11:42:32,059 (abs_task:1808) INFO: [train] Batch sampler: UnsortedBatchSampler(N-batch=28521, batch_size=256, key_file=exp_owsm/s2t_stats_raw_bpe50000/splits8/speech_shape/split.6, [cnode7-008:0/16] 2024-11-29 11:42:32,061 (abs_task:1809) INFO: [train] mini-batch sizes summary: N-batch=28521, mean=256.0, min=256, max=257 [2024-11-29 11:42:59,104] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:42:45,754] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:42:59,444] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:42:46,516] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:42:46,550] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:00,102] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:00,186] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:42:46,767] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:42:47,036] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:00,532] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:00,717] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:00,787] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:00,950] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:42:47,562] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:42:47,695] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:42:47,857] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:47,854] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:47,890] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:48,183] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:35,300] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:35,348] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:48,952] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:49,118] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:35,915] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:36,026] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:49,628] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:36,286] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:49,792] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:36,617] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:50,064] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:36,744] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:43:36,974] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:35,838] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:37,015] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:37,493] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:24,402] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:38,074] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:24,628] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:38,136] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:38,162] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:38,518] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:39,502] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:26,059] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:26,153] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:27,037] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:27,051] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:27,412] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:44:27,423] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:24,191] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:24,867] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:25,982] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:26,435] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:27,055] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:13,769] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:27,345] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:14,082] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:14,641] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:14,886] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:28,651] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:15,393] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:15,960] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:16,167] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:29,854] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:45:16,423] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [cnode7-008:0/16] 2024-11-29 11:46:16,436 (deepspeed_trainer:228) INFO: 9epoch:train:3701-3800batch: iter_time=2.391, loss_ctc=86.892, loss_att=68.797, acc=0.644, loss=74.216, grad_norm=3.643, loss_scale=1.000, learning_rate=1.393e-04, step_time=0.403 [cnode7-008:0/16] 2024-11-29 11:46:52,992 (deepspeed_trainer:228) INFO: 9epoch:train:3801-3900batch: iter_time=1.025e-04, loss_ctc=78.595, loss_att=58.489, acc=0.649, loss=64.527, grad_norm=3.345, loss_scale=1.000, learning_rate=1.392e-04, step_time=0.365 [2024-11-29 11:47:30,055] [INFO] [logging.py:129:log_dist] [Rank 0] step=124000, skipped=0, lr=[np.float64(0.00013912110775697278)], mom=[[0.9, 0.98]] [2024-11-29 11:47:30,059] [INFO] [timer.py:264:stop] epoch=0/micro_step=4000/global_step=4000, RunningAvgSamplesPerSec=43.53186566981625, CurrSamplesPerSec=40.74070058924704, MemAllocated=1.77GB, MaxMemAllocated=18.32GB [cnode7-008:0/16] 2024-11-29 11:47:30,060 (deepspeed_trainer:228) INFO: 9epoch:train:3901-4000batch: iter_time=9.855e-05, loss_ctc=84.906, loss_att=68.229, acc=0.638, loss=73.254, grad_norm=3.766, loss_scale=1.000, learning_rate=1.391e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 11:48:07,227 (deepspeed_trainer:228) INFO: 9epoch:train:4001-4100batch: iter_time=1.051e-04, loss_ctc=87.321, loss_att=71.850, acc=0.620, loss=76.511, grad_norm=3.882, loss_scale=1.000, learning_rate=1.391e-04, step_time=0.372 [cnode7-008:0/16] 2024-11-29 11:48:44,151 (deepspeed_trainer:228) INFO: 9epoch:train:4101-4200batch: iter_time=1.029e-04, loss_ctc=83.716, loss_att=65.456, acc=0.637, loss=70.926, grad_norm=3.477, loss_scale=1.000, learning_rate=1.390e-04, step_time=0.369 [cnode7-008:0/16] 2024-11-29 11:49:20,993 (deepspeed_trainer:228) INFO: 9epoch:train:4201-4300batch: iter_time=1.040e-04, loss_ctc=87.969, loss_att=75.395, acc=0.618, loss=79.145, grad_norm=3.884, loss_scale=1.000, learning_rate=1.390e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 11:49:57,936 (deepspeed_trainer:228) INFO: 9epoch:train:4301-4400batch: iter_time=1.033e-04, loss_ctc=88.198, loss_att=67.189, acc=0.641, loss=73.506, grad_norm=4.077, loss_scale=1.000, learning_rate=1.389e-04, step_time=0.369 [cnode7-008:0/16] 2024-11-29 11:50:35,069 (deepspeed_trainer:228) INFO: 9epoch:train:4401-4500batch: iter_time=1.043e-04, loss_ctc=97.796, loss_att=72.095, acc=0.644, loss=79.834, grad_norm=4.171, loss_scale=1.000, learning_rate=1.389e-04, step_time=0.371 [cnode7-008:0/16] 2024-11-29 11:51:12,240 (deepspeed_trainer:228) INFO: 9epoch:train:4501-4600batch: iter_time=1.040e-04, loss_ctc=95.778, loss_att=69.791, acc=0.655, loss=77.577, grad_norm=4.599, loss_scale=1.000, learning_rate=1.388e-04, step_time=0.371 [cnode7-008:0/16] 2024-11-29 11:51:51,857 (deepspeed_trainer:228) INFO: 9epoch:train:4601-4700batch: iter_time=1.118e-04, loss_ctc=86.951, loss_att=65.344, acc=0.637, loss=71.804, grad_norm=3.577, loss_scale=1.000, learning_rate=1.388e-04, step_time=0.396 [cnode7-008:0/16] 2024-11-29 11:52:30,219 (deepspeed_trainer:228) INFO: 9epoch:train:4701-4800batch: iter_time=1.186e-04, loss_ctc=82.489, loss_att=72.216, acc=0.648, loss=75.288, grad_norm=3.439, loss_scale=1.000, learning_rate=1.387e-04, step_time=0.383 [cnode7-008:0/16] 2024-11-29 11:53:09,727 (deepspeed_trainer:228) INFO: 9epoch:train:4801-4900batch: iter_time=1.125e-04, loss_ctc=84.477, loss_att=69.629, acc=0.638, loss=74.088, grad_norm=3.520, loss_scale=1.000, learning_rate=1.386e-04, step_time=0.395 [2024-11-29 11:53:46,843] [INFO] [logging.py:129:log_dist] [Rank 0] step=125000, skipped=0, lr=[np.float64(0.00013856351035257728)], mom=[[0.9, 0.98]] [2024-11-29 11:53:46,843] [INFO] [timer.py:264:stop] epoch=0/micro_step=5000/global_step=5000, RunningAvgSamplesPerSec=43.507529143602774, CurrSamplesPerSec=46.81749648264435, MemAllocated=1.77GB, MaxMemAllocated=18.32GB [cnode7-008:0/16] 2024-11-29 11:53:46,845 (deepspeed_trainer:228) INFO: 9epoch:train:4901-5000batch: iter_time=1.117e-04, loss_ctc=90.258, loss_att=69.426, acc=0.634, loss=75.703, grad_norm=3.685, loss_scale=1.000, learning_rate=1.386e-04, step_time=0.371 [cnode7-008:0/16] 2024-11-29 11:54:24,453 (deepspeed_trainer:228) INFO: 9epoch:train:5001-5100batch: iter_time=1.114e-04, loss_ctc=92.393, loss_att=73.177, acc=0.632, loss=78.925, grad_norm=4.012, loss_scale=1.000, learning_rate=1.385e-04, step_time=0.376 [cnode7-008:0/16] 2024-11-29 11:55:01,670 (deepspeed_trainer:228) INFO: 9epoch:train:5101-5200batch: iter_time=1.113e-04, loss_ctc=85.892, loss_att=71.896, acc=0.634, loss=76.111, grad_norm=3.661, loss_scale=1.000, learning_rate=1.385e-04, step_time=0.372 [cnode7-008:0/16] 2024-11-29 11:55:38,788 (deepspeed_trainer:228) INFO: 9epoch:train:5201-5300batch: iter_time=1.058e-04, loss_ctc=88.753, loss_att=67.507, acc=0.654, loss=73.871, grad_norm=3.712, loss_scale=1.000, learning_rate=1.384e-04, step_time=0.371 [cnode7-008:0/16] 2024-11-29 11:56:15,643 (deepspeed_trainer:228) INFO: 9epoch:train:5301-5400batch: iter_time=1.045e-04, loss_ctc=92.442, loss_att=77.815, acc=0.638, loss=82.204, grad_norm=3.731, loss_scale=1.000, learning_rate=1.384e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 11:56:52,461 (deepspeed_trainer:228) INFO: 9epoch:train:5401-5500batch: iter_time=1.033e-04, loss_ctc=87.832, loss_att=68.019, acc=0.643, loss=73.979, grad_norm=3.651, loss_scale=1.000, learning_rate=1.383e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 11:57:29,371 (deepspeed_trainer:228) INFO: 9epoch:train:5501-5600batch: iter_time=1.027e-04, loss_ctc=87.291, loss_att=74.017, acc=0.639, loss=78.008, grad_norm=3.563, loss_scale=1.000, learning_rate=1.383e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 11:57:42,358 (multiple_iter_factory:32) INFO: Building 3th iter-factory... [cnode7-008:0/16] 2024-11-29 11:58:09,074 (s2t:444) INFO: Optional Data Names: ('text_prev', 'text_ctc', 'text_spk2', 'text_spk3', 'text_spk4') [cnode7-008:0/16] 2024-11-29 11:58:14,576 (abs_task:1807) INFO: [train] dataset: ESPnetDataset( speech: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/wav.scp/split.2", "type": "kaldi_ark"} text_prev: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.prev/split.2", "type": "text"} text_ctc: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.ctc/split.2", "type": "text"} text: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text/split.2", "type": "text"} preprocess: ) [cnode7-008:0/16] 2024-11-29 11:58:14,576 (abs_task:1808) INFO: [train] Batch sampler: UnsortedBatchSampler(N-batch=28521, batch_size=256, key_file=exp_owsm/s2t_stats_raw_bpe50000/splits8/speech_shape/split.2, [cnode7-008:0/16] 2024-11-29 11:58:14,578 (abs_task:1809) INFO: [train] mini-batch sizes summary: N-batch=28521, mean=256.0, min=256, max=257 [2024-11-29 11:58:30,990] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:31,381] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:31,497] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:31,522] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:31,545] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:31,587] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:31,638] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:31,699] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:45,479] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:45,591] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:45,630] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:45,687] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:45,724] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:45,789] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:45,832] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:58:45,883] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:34,384] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:21,348] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:35,134] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:21,630] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:22,079] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:36,135] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:36,291] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:36,302] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:22,653] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:23,007] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:23,335] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:23,406] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:23,515] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:38,514] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:38,989] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 11:59:39,048] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:23,836] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:24,568] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:24,794] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:11,723] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:25,686] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:12,450] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:12,529] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:12,658] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:13,089] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:13,357] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:27,470] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:29,411] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:29,808] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:30,015] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:16,559] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:00:16,796] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:12,504] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:13,279] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:13,521] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:00,857] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:14,844] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:02,900] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:02,986] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:02,992] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:03,342] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:03,641] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:19,518] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:20,559] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:21,170] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:21,478] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:08,026] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:01:08,549] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [cnode7-008:0/16] 2024-11-29 12:02:17,403 (deepspeed_trainer:228) INFO: 9epoch:train:5601-5700batch: iter_time=2.462, loss_ctc=89.948, loss_att=66.971, acc=0.645, loss=73.869, grad_norm=3.640, loss_scale=1.000, learning_rate=1.382e-04, step_time=0.419 [cnode7-008:0/16] 2024-11-29 12:02:56,241 (deepspeed_trainer:228) INFO: 9epoch:train:5701-5800batch: iter_time=1.029e-04, loss_ctc=75.622, loss_att=58.344, acc=0.647, loss=63.535, grad_norm=3.502, loss_scale=1.000, learning_rate=1.381e-04, step_time=0.388 [cnode7-008:0/16] 2024-11-29 12:03:34,161 (deepspeed_trainer:228) INFO: 9epoch:train:5801-5900batch: iter_time=1.035e-04, loss_ctc=91.656, loss_att=69.658, acc=0.641, loss=76.277, grad_norm=3.931, loss_scale=1.000, learning_rate=1.381e-04, step_time=0.379 [2024-11-29 12:04:13,607] [INFO] [logging.py:129:log_dist] [Rank 0] step=126000, skipped=0, lr=[np.float64(0.00013801256420065193)], mom=[[0.9, 0.98]] [2024-11-29 12:04:13,608] [INFO] [timer.py:264:stop] epoch=0/micro_step=6000/global_step=6000, RunningAvgSamplesPerSec=43.419432490396815, CurrSamplesPerSec=20.2409165623665, MemAllocated=1.77GB, MaxMemAllocated=18.32GB [cnode7-008:0/16] 2024-11-29 12:04:13,609 (deepspeed_trainer:228) INFO: 9epoch:train:5901-6000batch: iter_time=1.044e-04, loss_ctc=78.427, loss_att=66.651, acc=0.636, loss=70.184, grad_norm=3.348, loss_scale=1.000, learning_rate=1.380e-04, step_time=0.390 [cnode7-008:0/16] 2024-11-29 12:04:50,225 (deepspeed_trainer:228) INFO: 9epoch:train:6001-6100batch: iter_time=1.082e-04, loss_ctc=81.082, loss_att=65.003, acc=0.632, loss=69.828, grad_norm=3.385, loss_scale=1.000, learning_rate=1.380e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 12:05:27,028 (deepspeed_trainer:228) INFO: 9epoch:train:6101-6200batch: iter_time=1.084e-04, loss_ctc=91.646, loss_att=75.439, acc=0.624, loss=80.316, grad_norm=3.846, loss_scale=1.000, learning_rate=1.379e-04, step_time=0.367 [cnode7-008:0/16] 2024-11-29 12:06:03,693 (deepspeed_trainer:228) INFO: 9epoch:train:6201-6300batch: iter_time=1.049e-04, loss_ctc=88.315, loss_att=67.481, acc=0.649, loss=73.720, grad_norm=4.003, loss_scale=1.000, learning_rate=1.379e-04, step_time=0.367 [cnode7-008:0/16] 2024-11-29 12:06:40,640 (deepspeed_trainer:228) INFO: 9epoch:train:6301-6400batch: iter_time=1.047e-04, loss_ctc=98.632, loss_att=72.063, acc=0.647, loss=80.022, grad_norm=3.811, loss_scale=1.000, learning_rate=1.378e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 12:07:17,447 (deepspeed_trainer:228) INFO: 9epoch:train:6401-6500batch: iter_time=1.066e-04, loss_ctc=90.971, loss_att=67.349, acc=0.652, loss=74.461, grad_norm=4.552, loss_scale=1.000, learning_rate=1.378e-04, step_time=0.369 [cnode7-008:0/16] 2024-11-29 12:07:54,346 (deepspeed_trainer:228) INFO: 9epoch:train:6501-6600batch: iter_time=1.055e-04, loss_ctc=84.110, loss_att=65.034, acc=0.651, loss=70.769, grad_norm=3.550, loss_scale=1.000, learning_rate=1.377e-04, step_time=0.369 [cnode7-008:0/16] 2024-11-29 12:08:31,192 (deepspeed_trainer:228) INFO: 9epoch:train:6601-6700batch: iter_time=1.052e-04, loss_ctc=85.542, loss_att=74.339, acc=0.640, loss=77.694, grad_norm=3.512, loss_scale=1.000, learning_rate=1.377e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 12:09:07,859 (deepspeed_trainer:228) INFO: 9epoch:train:6701-6800batch: iter_time=1.058e-04, loss_ctc=78.172, loss_att=65.172, acc=0.643, loss=69.084, grad_norm=3.276, loss_scale=1.000, learning_rate=1.376e-04, step_time=0.366 [cnode7-008:0/16] 2024-11-29 12:09:44,735 (deepspeed_trainer:228) INFO: 9epoch:train:6801-6900batch: iter_time=1.062e-04, loss_ctc=92.737, loss_att=72.964, acc=0.634, loss=78.874, grad_norm=3.757, loss_scale=1.000, learning_rate=1.375e-04, step_time=0.368 [2024-11-29 12:10:21,297] [INFO] [logging.py:129:log_dist] [Rank 0] step=127000, skipped=0, lr=[np.float64(0.00013746813811261736)], mom=[[0.9, 0.98]] [2024-11-29 12:10:21,297] [INFO] [timer.py:264:stop] epoch=0/micro_step=7000/global_step=7000, RunningAvgSamplesPerSec=43.568984018606386, CurrSamplesPerSec=45.96018432995493, MemAllocated=1.77GB, MaxMemAllocated=18.32GB [cnode7-008:0/16] 2024-11-29 12:10:21,299 (deepspeed_trainer:228) INFO: 9epoch:train:6901-7000batch: iter_time=1.053e-04, loss_ctc=87.207, loss_att=68.532, acc=0.631, loss=74.138, grad_norm=3.738, loss_scale=1.000, learning_rate=1.375e-04, step_time=0.365 [cnode7-008:0/16] 2024-11-29 12:10:58,237 (deepspeed_trainer:228) INFO: 9epoch:train:7001-7100batch: iter_time=1.040e-04, loss_ctc=91.427, loss_att=77.807, acc=0.634, loss=81.899, grad_norm=4.003, loss_scale=1.000, learning_rate=1.374e-04, step_time=0.369 [cnode7-008:0/16] 2024-11-29 12:11:35,313 (deepspeed_trainer:228) INFO: 9epoch:train:7101-7200batch: iter_time=1.049e-04, loss_ctc=95.119, loss_att=72.049, acc=0.651, loss=78.971, grad_norm=3.988, loss_scale=1.000, learning_rate=1.374e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 12:12:12,258 (deepspeed_trainer:228) INFO: 9epoch:train:7201-7300batch: iter_time=1.071e-04, loss_ctc=80.723, loss_att=67.154, acc=0.656, loss=71.230, grad_norm=3.556, loss_scale=1.000, learning_rate=1.373e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 12:12:50,866 (deepspeed_trainer:228) INFO: 9epoch:train:7301-7400batch: iter_time=1.049e-04, loss_ctc=95.466, loss_att=77.446, acc=0.634, loss=82.875, grad_norm=3.876, loss_scale=1.000, learning_rate=1.373e-04, step_time=0.385 [cnode7-008:0/16] 2024-11-29 12:13:29,103 (deepspeed_trainer:228) INFO: 9epoch:train:7401-7500batch: iter_time=1.121e-04, loss_ctc=81.648, loss_att=67.784, acc=0.649, loss=71.902, grad_norm=3.424, loss_scale=1.000, learning_rate=1.372e-04, step_time=0.382 [cnode7-008:0/16] 2024-11-29 12:13:32,380 (multiple_iter_factory:32) INFO: Building 4th iter-factory... [cnode7-008:0/16] 2024-11-29 12:13:59,520 (s2t:444) INFO: Optional Data Names: ('text_prev', 'text_ctc', 'text_spk2', 'text_spk3', 'text_spk4') [cnode7-008:0/16] 2024-11-29 12:14:05,159 (abs_task:1807) INFO: [train] dataset: ESPnetDataset( speech: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/wav.scp/split.1", "type": "kaldi_ark"} text_prev: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.prev/split.1", "type": "text"} text_ctc: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.ctc/split.1", "type": "text"} text: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text/split.1", "type": "text"} preprocess: ) [cnode7-008:0/16] 2024-11-29 12:14:05,159 (abs_task:1808) INFO: [train] Batch sampler: UnsortedBatchSampler(N-batch=28521, batch_size=256, key_file=exp_owsm/s2t_stats_raw_bpe50000/splits8/speech_shape/split.1, [cnode7-008:0/16] 2024-11-29 12:14:05,161 (abs_task:1809) INFO: [train] mini-batch sizes summary: N-batch=28521, mean=256.0, min=256, max=257 [2024-11-29 12:14:18,588] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:18,742] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:33,050] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:33,101] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:33,174] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:33,541] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:33,607] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:19,741] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:33,914] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:19,999] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:34,102] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:20,103] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:20,119] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:34,189] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:20,351] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:14:20,496] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:07,238] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:21,774] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:08,268] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:22,625] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:08,866] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:23,158] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:23,231] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:23,258] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:09,392] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:09,445] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:23,618] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:09,853] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:10,026] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:11,763] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:26,334] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:26,683] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:56,220] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:10,671] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:10,845] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:11,443] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:57,643] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:11,800] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:58,001] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:12,072] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:58,446] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:58,951] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:59,123] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:15:59,376] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:15,052] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:02,757] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:17,359] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:17,818] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:58,378] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:59,040] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:59,995] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:46,244] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:17:00,402] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:17:00,712] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:47,502] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:47,584] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:47,947] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:48,279] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:48,706] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:48,875] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:17:05,076] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:16:52,153] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:17:06,908] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:17:07,593] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [cnode7-008:0/16] 2024-11-29 12:18:11,523 (deepspeed_trainer:228) INFO: 9epoch:train:7501-7600batch: iter_time=2.405, loss_ctc=85.570, loss_att=61.606, acc=0.644, loss=68.792, grad_norm=3.596, loss_scale=1.000, learning_rate=1.372e-04, step_time=0.419 [cnode7-008:0/16] 2024-11-29 12:18:48,133 (deepspeed_trainer:228) INFO: 9epoch:train:7601-7700batch: iter_time=1.110e-04, loss_ctc=81.565, loss_att=65.304, acc=0.631, loss=70.174, grad_norm=3.704, loss_scale=1.000, learning_rate=1.371e-04, step_time=0.366 [cnode7-008:0/16] 2024-11-29 12:19:24,659 (deepspeed_trainer:228) INFO: 9epoch:train:7701-7800batch: iter_time=1.078e-04, loss_ctc=82.878, loss_att=63.937, acc=0.641, loss=69.606, grad_norm=3.732, loss_scale=1.000, learning_rate=1.371e-04, step_time=0.365 [cnode7-008:0/16] 2024-11-29 12:20:01,735 (deepspeed_trainer:228) INFO: 9epoch:train:7801-7900batch: iter_time=1.075e-04, loss_ctc=82.307, loss_att=66.161, acc=0.648, loss=71.001, grad_norm=3.383, loss_scale=1.000, learning_rate=1.370e-04, step_time=0.370 [2024-11-29 12:20:38,523] [INFO] [logging.py:129:log_dist] [Rank 0] step=128000, skipped=0, lr=[np.float64(0.00013693010449411554)], mom=[[0.9, 0.98]] [2024-11-29 12:20:38,524] [INFO] [timer.py:264:stop] epoch=0/micro_step=8000/global_step=8000, RunningAvgSamplesPerSec=43.548777155487606, CurrSamplesPerSec=42.01566948624144, MemAllocated=1.77GB, MaxMemAllocated=18.32GB [cnode7-008:0/16] 2024-11-29 12:20:38,526 (deepspeed_trainer:228) INFO: 9epoch:train:7901-8000batch: iter_time=1.064e-04, loss_ctc=82.102, loss_att=76.118, acc=0.605, loss=77.922, grad_norm=3.711, loss_scale=1.000, learning_rate=1.370e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 12:21:15,193 (deepspeed_trainer:228) INFO: 9epoch:train:8001-8100batch: iter_time=1.091e-04, loss_ctc=85.831, loss_att=65.026, acc=0.628, loss=71.280, grad_norm=4.107, loss_scale=1.000, learning_rate=1.369e-04, step_time=0.367 [cnode7-008:0/16] 2024-11-29 12:21:51,958 (deepspeed_trainer:228) INFO: 9epoch:train:8101-8200batch: iter_time=1.079e-04, loss_ctc=89.311, loss_att=70.186, acc=0.648, loss=75.915, grad_norm=3.798, loss_scale=1.000, learning_rate=1.368e-04, step_time=0.367 [cnode7-008:0/16] 2024-11-29 12:22:28,895 (deepspeed_trainer:228) INFO: 9epoch:train:8201-8300batch: iter_time=1.080e-04, loss_ctc=105.340, loss_att=71.749, acc=0.646, loss=81.828, grad_norm=5.428, loss_scale=1.000, learning_rate=1.368e-04, step_time=0.369 [cnode7-008:0/16] 2024-11-29 12:23:05,730 (deepspeed_trainer:228) INFO: 9epoch:train:8301-8400batch: iter_time=1.074e-04, loss_ctc=92.744, loss_att=69.476, acc=0.632, loss=76.446, grad_norm=4.164, loss_scale=1.000, learning_rate=1.367e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 12:23:42,730 (deepspeed_trainer:228) INFO: 9epoch:train:8401-8500batch: iter_time=1.065e-04, loss_ctc=81.543, loss_att=65.761, acc=0.651, loss=70.496, grad_norm=3.599, loss_scale=1.000, learning_rate=1.367e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 12:24:21,164 (deepspeed_trainer:228) INFO: 9epoch:train:8501-8600batch: iter_time=1.043e-04, loss_ctc=82.444, loss_att=68.672, acc=0.634, loss=72.838, grad_norm=3.684, loss_scale=1.000, learning_rate=1.366e-04, step_time=0.384 [cnode7-008:0/16] 2024-11-29 12:24:58,894 (deepspeed_trainer:228) INFO: 9epoch:train:8601-8700batch: iter_time=1.019e-04, loss_ctc=81.574, loss_att=67.831, acc=0.643, loss=71.942, grad_norm=3.412, loss_scale=1.000, learning_rate=1.366e-04, step_time=0.377 [cnode7-008:0/16] 2024-11-29 12:25:35,878 (deepspeed_trainer:228) INFO: 9epoch:train:8701-8800batch: iter_time=1.063e-04, loss_ctc=88.018, loss_att=69.554, acc=0.627, loss=75.127, grad_norm=3.493, loss_scale=1.000, learning_rate=1.365e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 12:26:15,158 (deepspeed_trainer:228) INFO: 9epoch:train:8801-8900batch: iter_time=1.078e-04, loss_ctc=90.870, loss_att=73.213, acc=0.633, loss=78.534, grad_norm=3.814, loss_scale=1.000, learning_rate=1.365e-04, step_time=0.392 [2024-11-29 12:26:52,757] [INFO] [logging.py:129:log_dist] [Rank 0] step=129000, skipped=0, lr=[np.float64(0.00013639833921938588)], mom=[[0.9, 0.98]] [2024-11-29 12:26:52,757] [INFO] [timer.py:264:stop] epoch=0/micro_step=9000/global_step=9000, RunningAvgSamplesPerSec=43.565171953517265, CurrSamplesPerSec=40.2545576892125, MemAllocated=1.77GB, MaxMemAllocated=18.32GB [cnode7-008:0/16] 2024-11-29 12:26:52,759 (deepspeed_trainer:228) INFO: 9epoch:train:8901-9000batch: iter_time=1.074e-04, loss_ctc=87.918, loss_att=69.878, acc=0.645, loss=75.297, grad_norm=3.961, loss_scale=1.000, learning_rate=1.364e-04, step_time=0.376 [cnode7-008:0/16] 2024-11-29 12:27:29,573 (deepspeed_trainer:228) INFO: 9epoch:train:9001-9100batch: iter_time=1.200e-04, loss_ctc=94.141, loss_att=76.552, acc=0.638, loss=81.792, grad_norm=3.840, loss_scale=1.000, learning_rate=1.364e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 12:28:06,408 (deepspeed_trainer:228) INFO: 9epoch:train:9101-9200batch: iter_time=1.181e-04, loss_ctc=84.934, loss_att=66.323, acc=0.654, loss=71.889, grad_norm=3.651, loss_scale=1.000, learning_rate=1.363e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 12:28:43,117 (deepspeed_trainer:228) INFO: 9epoch:train:9201-9300batch: iter_time=1.114e-04, loss_ctc=85.529, loss_att=70.720, acc=0.634, loss=75.158, grad_norm=3.505, loss_scale=1.000, learning_rate=1.363e-04, step_time=0.367 [cnode7-008:0/16] 2024-11-29 12:29:14,254 (multiple_iter_factory:32) INFO: Building 5th iter-factory... [cnode7-008:0/16] 2024-11-29 12:29:41,814 (s2t:444) INFO: Optional Data Names: ('text_prev', 'text_ctc', 'text_spk2', 'text_spk3', 'text_spk4') [cnode7-008:0/16] 2024-11-29 12:29:47,265 (abs_task:1807) INFO: [train] dataset: ESPnetDataset( speech: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/wav.scp/split.3", "type": "kaldi_ark"} text_prev: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.prev/split.3", "type": "text"} text_ctc: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.ctc/split.3", "type": "text"} text: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text/split.3", "type": "text"} preprocess: ) [cnode7-008:0/16] 2024-11-29 12:29:47,265 (abs_task:1808) INFO: [train] Batch sampler: UnsortedBatchSampler(N-batch=28521, batch_size=256, key_file=exp_owsm/s2t_stats_raw_bpe50000/splits8/speech_shape/split.3, [cnode7-008:0/16] 2024-11-29 12:29:47,268 (abs_task:1809) INFO: [train] mini-batch sizes summary: N-batch=28521, mean=256.0, min=256, max=257 [2024-11-29 12:30:12,564] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:03,050] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:13,936] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:03,453] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:14,016] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:03,579] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:14,379] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:14,953] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:04,443] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:15,418] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:15,450] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:04,924] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:04,974] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:15,603] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:05,189] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:05,210] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:01,485] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:02,053] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:02,573] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:03,200] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:03,822] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:03,873] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:53,333] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:53,738] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:04,327] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:54,367] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:54,866] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:54,987] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:55,163] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:55,218] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:30:55,600] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:07,837] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:50,450] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:50,702] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:50,934] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:51,990] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:52,598] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:52,744] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:52,946] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:42,417] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:42,972] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:43,262] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:44,492] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:44,543] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:44,989] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:45,179] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:45,549] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:31:59,185] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:38,286] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:38,371] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:38,684] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:39,749] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:40,553] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:30,299] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:41,660] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:32,504] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:32,860] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:43,735] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:34,450] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:34,563] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:35,162] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:35,260] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:35,262] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:32:48,458] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [cnode7-008:0/16] 2024-11-29 12:33:24,726 (deepspeed_trainer:228) INFO: 9epoch:train:9301-9400batch: iter_time=2.450, loss_ctc=82.882, loss_att=65.558, acc=0.654, loss=70.739, grad_norm=3.529, loss_scale=1.000, learning_rate=1.362e-04, step_time=0.365 [cnode7-008:0/16] 2024-11-29 12:34:01,751 (deepspeed_trainer:228) INFO: 9epoch:train:9401-9500batch: iter_time=1.079e-04, loss_ctc=84.153, loss_att=61.718, acc=0.644, loss=68.444, grad_norm=3.507, loss_scale=1.000, learning_rate=1.362e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 12:34:39,508 (deepspeed_trainer:228) INFO: 9epoch:train:9501-9600batch: iter_time=1.351e-04, loss_ctc=82.121, loss_att=63.981, acc=0.641, loss=69.424, grad_norm=3.780, loss_scale=1.000, learning_rate=1.361e-04, step_time=0.377 [cnode7-008:0/16] 2024-11-29 12:35:16,974 (deepspeed_trainer:228) INFO: 9epoch:train:9601-9700batch: iter_time=1.362e-04, loss_ctc=83.668, loss_att=64.258, acc=0.629, loss=70.054, grad_norm=4.025, loss_scale=1.000, learning_rate=1.361e-04, step_time=0.374 [cnode7-008:0/16] 2024-11-29 12:35:55,482 (deepspeed_trainer:228) INFO: 9epoch:train:9701-9800batch: iter_time=1.075e-04, loss_ctc=84.034, loss_att=69.316, acc=0.645, loss=73.741, grad_norm=3.629, loss_scale=1.000, learning_rate=1.360e-04, step_time=0.384 [cnode7-008:0/16] 2024-11-29 12:36:33,798 (deepspeed_trainer:228) INFO: 9epoch:train:9801-9900batch: iter_time=1.080e-04, loss_ctc=80.207, loss_att=68.455, acc=0.619, loss=71.975, grad_norm=3.601, loss_scale=1.000, learning_rate=1.360e-04, step_time=0.383 [2024-11-29 12:37:11,173] [INFO] [logging.py:129:log_dist] [Rank 0] step=130000, skipped=0, lr=[np.float64(0.00013587272151096603)], mom=[[0.9, 0.98]] [2024-11-29 12:37:11,173] [INFO] [timer.py:264:stop] epoch=0/micro_step=10000/global_step=10000, RunningAvgSamplesPerSec=43.59374296355244, CurrSamplesPerSec=43.23946191589301, MemAllocated=1.77GB, MaxMemAllocated=18.32GB [cnode7-008:0/16] 2024-11-29 12:37:11,175 (deepspeed_trainer:228) INFO: 9epoch:train:9901-10000batch: iter_time=1.106e-04, loss_ctc=85.051, loss_att=63.829, acc=0.638, loss=70.175, grad_norm=3.991, loss_scale=1.000, learning_rate=1.359e-04, step_time=0.373 [cnode7-008:0/16] 2024-11-29 12:37:50,366 (deepspeed_trainer:228) INFO: 9epoch:train:10001-10100batch: iter_time=1.099e-04, loss_ctc=89.351, loss_att=68.220, acc=0.645, loss=74.551, grad_norm=4.106, loss_scale=1.000, learning_rate=1.358e-04, step_time=0.392 [cnode7-008:0/16] 2024-11-29 12:38:29,106 (deepspeed_trainer:228) INFO: 9epoch:train:10101-10200batch: iter_time=1.085e-04, loss_ctc=104.571, loss_att=74.283, acc=0.651, loss=83.355, grad_norm=5.184, loss_scale=1.000, learning_rate=1.358e-04, step_time=0.387 [cnode7-008:0/16] 2024-11-29 12:39:05,897 (deepspeed_trainer:228) INFO: 9epoch:train:10201-10300batch: iter_time=1.068e-04, loss_ctc=86.078, loss_att=62.744, acc=0.639, loss=69.765, grad_norm=3.707, loss_scale=1.000, learning_rate=1.357e-04, step_time=0.367 [cnode7-008:0/16] 2024-11-29 12:39:42,443 (deepspeed_trainer:228) INFO: 9epoch:train:10301-10400batch: iter_time=1.067e-04, loss_ctc=77.817, loss_att=65.348, acc=0.654, loss=69.049, grad_norm=3.481, loss_scale=1.000, learning_rate=1.357e-04, step_time=0.365 [cnode7-008:0/16] 2024-11-29 12:40:19,015 (deepspeed_trainer:228) INFO: 9epoch:train:10401-10500batch: iter_time=1.091e-04, loss_ctc=83.464, loss_att=70.344, acc=0.638, loss=74.294, grad_norm=3.633, loss_scale=1.000, learning_rate=1.356e-04, step_time=0.365 [cnode7-008:0/16] 2024-11-29 12:40:55,936 (deepspeed_trainer:228) INFO: 9epoch:train:10501-10600batch: iter_time=1.088e-04, loss_ctc=86.931, loss_att=68.986, acc=0.632, loss=74.356, grad_norm=4.135, loss_scale=1.000, learning_rate=1.356e-04, step_time=0.369 [cnode7-008:0/16] 2024-11-29 12:41:32,833 (deepspeed_trainer:228) INFO: 9epoch:train:10601-10700batch: iter_time=1.091e-04, loss_ctc=88.329, loss_att=69.883, acc=0.639, loss=75.422, grad_norm=3.879, loss_scale=1.000, learning_rate=1.355e-04, step_time=0.369 [cnode7-008:0/16] 2024-11-29 12:42:09,695 (deepspeed_trainer:228) INFO: 9epoch:train:10701-10800batch: iter_time=1.083e-04, loss_ctc=85.162, loss_att=71.297, acc=0.636, loss=75.489, grad_norm=3.860, loss_scale=1.000, learning_rate=1.355e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 12:42:46,985 (deepspeed_trainer:228) INFO: 9epoch:train:10801-10900batch: iter_time=1.064e-04, loss_ctc=92.453, loss_att=69.334, acc=0.654, loss=76.217, grad_norm=4.122, loss_scale=1.000, learning_rate=1.354e-04, step_time=0.373 [2024-11-29 12:43:24,157] [INFO] [logging.py:129:log_dist] [Rank 0] step=131000, skipped=0, lr=[np.float64(0.00013535313382445377)], mom=[[0.9, 0.98]] [2024-11-29 12:43:24,158] [INFO] [timer.py:264:stop] epoch=0/micro_step=11000/global_step=11000, RunningAvgSamplesPerSec=43.61694259897927, CurrSamplesPerSec=42.74748911961467, MemAllocated=1.77GB, MaxMemAllocated=18.32GB [cnode7-008:0/16] 2024-11-29 12:43:24,159 (deepspeed_trainer:228) INFO: 9epoch:train:10901-11000batch: iter_time=1.086e-04, loss_ctc=88.252, loss_att=70.379, acc=0.647, loss=75.695, grad_norm=3.846, loss_scale=1.000, learning_rate=1.354e-04, step_time=0.371 [cnode7-008:0/16] 2024-11-29 12:44:01,376 (deepspeed_trainer:228) INFO: 9epoch:train:11001-11100batch: iter_time=1.070e-04, loss_ctc=87.389, loss_att=67.934, acc=0.646, loss=73.791, grad_norm=3.618, loss_scale=1.000, learning_rate=1.353e-04, step_time=0.372 [cnode7-008:0/16] 2024-11-29 12:44:38,607 (deepspeed_trainer:228) INFO: 9epoch:train:11101-11200batch: iter_time=1.079e-04, loss_ctc=86.183, loss_att=70.814, acc=0.640, loss=75.376, grad_norm=3.648, loss_scale=1.000, learning_rate=1.353e-04, step_time=0.372 [cnode7-008:0/16] 2024-11-29 12:45:00,969 (multiple_iter_factory:32) INFO: Building 6th iter-factory... [cnode7-008:0/16] 2024-11-29 12:45:29,265 (s2t:444) INFO: Optional Data Names: ('text_prev', 'text_ctc', 'text_spk2', 'text_spk3', 'text_spk4') [cnode7-008:0/16] 2024-11-29 12:45:34,624 (abs_task:1807) INFO: [train] dataset: ESPnetDataset( speech: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/wav.scp/split.7", "type": "kaldi_ark"} text_prev: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.prev/split.7", "type": "text"} text_ctc: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text.ctc/split.7", "type": "text"} text: {"path": "exp_owsm/s2t_stats_raw_bpe50000/splits8/text/split.7", "type": "text"} preprocess: ) [cnode7-008:0/16] 2024-11-29 12:45:34,625 (abs_task:1808) INFO: [train] Batch sampler: UnsortedBatchSampler(N-batch=28521, batch_size=256, key_file=exp_owsm/s2t_stats_raw_bpe50000/splits8/speech_shape/split.7, [cnode7-008:0/16] 2024-11-29 12:45:34,627 (abs_task:1809) INFO: [train] mini-batch sizes summary: N-batch=28521, mean=256.0, min=256, max=257 [2024-11-29 12:45:50,543] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:01,488] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:01,667] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:45:51,050] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:01,948] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:45:51,288] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:45:51,353] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:02,232] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:45:51,409] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:45:51,579] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:45:51,634] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:45:51,702] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:02,806] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:02,997] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:03,056] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:03,100] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:49,967] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:39,198] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:50,375] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:50,743] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:40,045] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:51,016] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:51,081] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:51,519] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:51,625] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:40,945] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:41,819] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:41,854] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:41,956] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:53,812] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:45,318] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:46:45,353] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:38,117] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:38,677] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:28,018] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:39,016] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:39,894] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:39,987] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:29,500] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:40,398] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:40,435] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:29,738] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:30,681] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:31,263] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:31,799] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:43,801] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:37,118] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:47:37,365] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:26,806] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:16,610] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:27,507] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:27,678] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:28,202] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:17,551] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:28,544] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:17,797] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:29,411] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:29,597] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:18,917] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:20,647] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:33,597] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:24,565] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:27,752] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [2024-11-29 12:48:28,166] [INFO] [real_accelerator.py:219:get_accelerator] Setting ds_accelerator to cuda (auto detect) [cnode7-008:0/16] 2024-11-29 12:49:24,791 (deepspeed_trainer:228) INFO: 9epoch:train:11201-11300batch: iter_time=2.412, loss_ctc=86.044, loss_att=65.620, acc=0.652, loss=71.744, grad_norm=3.689, loss_scale=1.000, learning_rate=1.352e-04, step_time=0.447 [cnode7-008:0/16] 2024-11-29 12:50:03,960 (deepspeed_trainer:228) INFO: 9epoch:train:11301-11400batch: iter_time=1.027e-04, loss_ctc=77.317, loss_att=56.088, acc=0.654, loss=62.447, grad_norm=3.481, loss_scale=1.000, learning_rate=1.352e-04, step_time=0.395 [cnode7-008:0/16] 2024-11-29 12:50:41,713 (deepspeed_trainer:228) INFO: 9epoch:train:11401-11500batch: iter_time=1.018e-04, loss_ctc=82.576, loss_att=65.466, acc=0.643, loss=70.588, grad_norm=3.700, loss_scale=1.000, learning_rate=1.351e-04, step_time=0.377 [cnode7-008:0/16] 2024-11-29 12:51:18,516 (deepspeed_trainer:228) INFO: 9epoch:train:11501-11600batch: iter_time=1.041e-04, loss_ctc=85.113, loss_att=67.906, acc=0.624, loss=73.065, grad_norm=4.136, loss_scale=1.000, learning_rate=1.351e-04, step_time=0.368 [cnode7-008:0/16] 2024-11-29 12:51:55,525 (deepspeed_trainer:228) INFO: 9epoch:train:11601-11700batch: iter_time=1.071e-04, loss_ctc=82.448, loss_att=63.138, acc=0.647, loss=68.936, grad_norm=3.474, loss_scale=1.000, learning_rate=1.350e-04, step_time=0.370 [cnode7-008:0/16] 2024-11-29 12:52:32,507 (deepspeed_trainer:228) INFO: 9epoch:train:11701-11800batch: iter_time=1.080e-04, loss_ctc=84.744, loss_att=73.145, acc=0.618, loss=76.647, grad_norm=3.925, loss_scale=1.000, learning_rate=1.350e-04, step_time=0.369 [cnode7-008:0/16] 2024-11-29 12:53:09,562 (deepspeed_trainer:228) INFO: 9epoch:train:11801-11900batch: iter_time=1.090e-04, loss_ctc=85.651, loss_att=63.996, acc=0.647, loss=70.504, grad_norm=3.957, loss_scale=1.000, learning_rate=1.349e-04, step_time=0.370 Process SpawnProcess-5: Traceback (most recent call last): File "/mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap self.run() File "/mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/multiprocessing/process.py", line 108, in run self._target(*self._args, **self._kwargs) File "/mnt/home/williamchen/espnet/espnet2/tasks/abs_task.py", line 1568, in main_worker cls.trainer.run( File "/mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py", line 132, in run cls.train_one_epoch( File "/mnt/home/williamchen/espnet/espnet2/train/deepspeed_trainer.py", line 204, in train_one_epoch loss, stats, weight = model(**batch) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, **kwargs) File "/mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 18, in wrapped_fn ret_val = func(*args, **kwargs) File "/mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 1899, in forward loss = self.module(*inputs, **kwargs) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, **kwargs) File "/mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py", line 225, in forward loss_att, acc_att, cer_att, wer_att = self._calc_att_loss( File "/mnt/home/williamchen/espnet/espnet2/s2t/espnet_model.py", line 391, in _calc_att_loss decoder_out, _ = self.decoder( File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, **kwargs) File "/mnt/home/williamchen/espnet/espnet2/asr/decoder/transformer_decoder.py", line 155, in forward x, tgt_mask, memory, memory_mask = checkpoint(decoder_layer, File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 600, in _fn return fn(*args, **kwargs) File "/mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/activation_checkpointing/checkpointing.py", line 951, in checkpoint CheckpointFunction.apply(function, all_outputs, *args) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/autograd/function.py", line 574, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File "/mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/site-packages/deepspeed/runtime/activation_checkpointing/checkpointing.py", line 542, in forward outputs = run_function(*inputs_cuda) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, **kwargs) File "/mnt/home/williamchen/espnet/espnet/nets/pytorch_backend/transformer/decoder_layer.py", line 170, in forward x = residual + self.dropout(self.feed_forward(x)) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, **kwargs) File "/mnt/home/williamchen/espnet/espnet/nets/pytorch_backend/transformer/positionwise_feed_forward.py", line 32, in forward return self.w_2(self.dropout(self.activation(self.w_1(x)))) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl return forward_call(*args, **kwargs) File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/nn/modules/linear.py", line 117, in forward return F.linear(input, self.weight, self.bias) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 6.91 GiB. GPU 4 has a total capacity of 79.11 GiB of which 5.21 GiB is free. Including non-PyTorch memory, this process has 73.89 GiB memory in use. Of the allocated memory 66.78 GiB is allocated by PyTorch, and 1.08 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables) cnode7-008:3116459:3126114 [0] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116459:3117412 [0] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116464:3126120 [5] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116464:3117408 [5] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116461:3126110 [2] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116461:3117411 [2] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116462:3126112 [3] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116462:3117407 [3] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116465:3126111 [6] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116465:3117405 [6] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116466:3126122 [7] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116466:3117406 [7] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116460:3126113 [1] NCCL INFO [Service thread] Connection closed by localRank 4 cnode7-008:3116460:3117410 [1] NCCL INFO [Service thread] Connection closed by localRank 4 W1129 12:53:53.284000 23456244418368 torch/multiprocessing/spawn.py:146] Terminating process 3116459 via signal SIGTERM W1129 12:53:53.285000 23456244418368 torch/multiprocessing/spawn.py:146] Terminating process 3116460 via signal SIGTERM W1129 12:53:53.285000 23456244418368 torch/multiprocessing/spawn.py:146] Terminating process 3116461 via signal SIGTERM W1129 12:53:53.286000 23456244418368 torch/multiprocessing/spawn.py:146] Terminating process 3116462 via signal SIGTERM W1129 12:53:53.287000 23456244418368 torch/multiprocessing/spawn.py:146] Terminating process 3116464 via signal SIGTERM W1129 12:53:53.287000 23456244418368 torch/multiprocessing/spawn.py:146] Terminating process 3116465 via signal SIGTERM W1129 12:53:53.288000 23456244418368 torch/multiprocessing/spawn.py:146] Terminating process 3116466 via signal SIGTERM Traceback (most recent call last): File "/mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "/mnt/home/williamchen/espnet/tools/miniconda/envs/espnet/lib/python3.10/runpy.py", line 86, in _run_code exec(code, run_globals) File "/mnt/home/williamchen/espnet/espnet2/bin/s2t_train.py", line 23, in main() File "/mnt/home/williamchen/espnet/espnet2/bin/s2t_train.py", line 19, in main S2TTask.main(cmd=cmd) File "/mnt/home/williamchen/espnet/espnet2/tasks/abs_task.py", line 1273, in main while not ProcessContext(processes, error_files).join(): File "/mnt/home/williamchen/.local/lib/python3.10/site-packages/torch/multiprocessing/spawn.py", line 178, in join raise ProcessExitedException( torch.multiprocessing.spawn.ProcessExitedException: process 4 terminated with exit code 1 srun: error: cnode7-008: task 0: Exited with exit code 1 srun: Job step aborted: Waiting up to 32 seconds for job step to finish.