V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] Output code: 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # AOT ID: ['0_inference']
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from ctypes import c_void_p, c_long
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import torch
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import random
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import os
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import tempfile
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from math import inf, nan
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.hooks import run_intermediate_hooks
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.utils import maybe_profile
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.codegen.memory_planning import _align as align
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch import device, empty_strided
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.async_compile import AsyncCompile
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.select_algorithm import extern_kernels
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.codegen.multi_kernel import MultiKernelCall
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] aten = torch.ops.aten
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_ops = torch.ops.inductor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] _quantized = torch.ops._quantized
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride = torch._C._dynamo.guards.assert_size_stride
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] empty_strided_cpu = torch._C._dynamo.guards._empty_strided_cpu
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] empty_strided_cuda = torch._C._dynamo.guards._empty_strided_cuda
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] reinterpret_tensor = torch._C._dynamo.guards._reinterpret_tensor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] alloc_from_pool = torch.ops.inductor._alloc_from_pool
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] async_compile = AsyncCompile()
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/kx/ckxvjo3hhiznc4ieahqn6ez5oq2jhe6or6jnoamh6yry3pp64dc5.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [emb_3], Original ATen: [aten.cat]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # emb_3 => cat
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_cat_0 = async_compile.triton('triton_poi_fused_cat_0', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[512], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*i64', 1: '*fp32', 2: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_cat_0', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 1, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_cat_0(in_ptr0, out_ptr0, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 512
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = xindex < xnumel
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 256
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp5 = tl.load(in_ptr0 + (0))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp6 = tl.broadcast_to(tmp5, [XBLOCK])
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = x0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.full([1], 0, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 >= tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tl.full([1], 128, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp4 = tmp0 < tmp3
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7 = tmp6.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp8 = tmp0.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9 = -9.210340371976184
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp10 = tmp8 * tmp9
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp11 = 0.0078125
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp12 = tmp10 * tmp11
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp13 = tl_math.exp(tmp12)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp14 = tmp7 * tmp13
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp15 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp16 = tmp14 * tmp15
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp17 = tl_math.sin(tmp16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp18 = tl.full(tmp17.shape, 0.0, tmp17.dtype)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19 = tl.where(tmp4, tmp17, tmp18)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp20 = tmp0 >= tmp3
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp21 = tl.full([1], 256, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp22 = tmp0 < tmp21
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp23 = (-128) + x0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp24 = tmp23.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp25 = tmp24 * tmp9
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp26 = tmp25 * tmp11
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp27 = tl_math.exp(tmp26)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp28 = tmp7 * tmp27
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp29 = tmp28 * tmp15
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp30 = tl_math.cos(tmp29)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp31 = tl.full(tmp30.shape, 0.0, tmp30.dtype)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp32 = tl.where(tmp20, tmp30, tmp31)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp33 = tl.where(tmp4, tmp19, tmp32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x2), tmp33, xmask)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_heuristics import grid, split_scan_grid, grid_combo_kernels, start_graph, end_graph
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._C import _cuda_getCurrentRawStream as get_raw_stream
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/ay/caykwk6ibs4q2wisgu3nlt2wrwimvicwlfsqqsfonccmq3gwcupp.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [emb_4, sample, to_2], Original ATen: [aten._to_copy, aten.cat, aten.constant_pad_nd, aten.view]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # emb_4 => cat_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # sample => constant_pad_nd, view_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # to_2 => convert_element_type_3
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__to_copy_cat_constant_pad_nd_view_1 = async_compile.triton('triton_poi_fused__to_copy_cat_constant_pad_nd_view_1', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[2048], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*fp32', 1: '*bf16', 2: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused__to_copy_cat_constant_pad_nd_view_1', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused__to_copy_cat_constant_pad_nd_view_1(in_ptr0, out_ptr0, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 2048
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 1024
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x1 = (xindex // 1024)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = x0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.full([1], 256, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 < tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tl.full([1], 0, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp4 = tmp0 >= tmp3
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp5 = tl.full([1], 128, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp6 = tmp0 < tmp5
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7 = tmp6 & tmp2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp8 = tl.load(in_ptr0 + (128 + (256*x1) + x0), tmp7, eviction_policy='evict_last', other=0.0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9 = tmp0 >= tmp5
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp10 = tmp9 & tmp2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp11 = tl.load(in_ptr0 + ((256*x1) + ((-128) + x0)), tmp10, eviction_policy='evict_last', other=0.0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp12 = tl.where(tmp6, tmp8, tmp11)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp13 = tmp12.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp14 = tl.full(tmp13.shape, 0.0, tmp13.dtype)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp15 = tl.where(tmp2, tmp13, tmp14)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x2), tmp15, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/zz/czzffj47sqhn5iimhr4px5p266u5c2wpnfsrtuofxf6hak6q2da6.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [sample, sample_1, sample_2], Original ATen: [aten.add, aten.constant_pad_nd, aten.silu, aten.view]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # sample => add_2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # sample_1 => convert_element_type_4, convert_element_type_5, mul_5, sigmoid
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # sample_2 => constant_pad_nd_1, view_5
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_constant_pad_nd_silu_view_2 = async_compile.triton('triton_poi_fused_add_constant_pad_nd_silu_view_2', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[4096], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_add_constant_pad_nd_silu_view_2', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_add_constant_pad_nd_silu_view_2(in_ptr0, in_ptr1, out_ptr0, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 4096
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 2048
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x1 = (xindex // 2048)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = x0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.full([1], 1152, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 < tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tl.load(in_ptr0 + (x0 + (1152*x1)), tmp2, other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp4 = tl.load(in_ptr1 + (x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp5 = tmp3 + tmp4
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp6 = tmp5.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7 = tl.sigmoid(tmp6)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp8 = tmp6 * tmp7
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9 = tmp8.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp10 = tl.full(tmp9.shape, 0.0, tmp9.dtype)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp11 = tl.where(tmp2, tmp9, tmp10)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x2), tmp11, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/kz/ckzvxx7soy6esfx6au7rln5h7czwx6pg24olgmfwjfogdbx33xfn.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [latent], Original ATen: [aten.convolution]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # latent => convolution
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_convolution_3 = async_compile.triton('triton_poi_fused_convolution_3', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[8, 16384], tile_hint=TileHint.SQUARE,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: 'i32', 3: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 3), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_convolution_3', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 1, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_convolution_3(in_ptr0, out_ptr0, ynumel, xnumel, YBLOCK : tl.constexpr, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     ynumel = 8
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 16384
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     yoffset = tl.program_id(1) * YBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     yindex = yoffset + tl.arange(0, YBLOCK)[None, :]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     ymask = yindex < ynumel
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK, YBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     y3 = yindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     y0 = yindex % 4
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     y1 = (yindex // 4)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = tl.load(in_ptr0 + (x2 + (16384*y3)), ymask, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (y0 + (4*x2) + (65536*y1)), tmp0, ymask)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/wt/cwt7sxdkktlnm3e4lgnzgzqerira5sdkczyq6upiuxa7c35e5vqu.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [latent], Original ATen: [aten.convolution]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # latent => convolution
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_tem_fused_convolution_4 = async_compile.triton('triton_tem_fused_convolution_4', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.template(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     num_stages=2,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     num_warps=4,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'kernel_name': 'triton_tem_fused_convolution_4', 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_tem_fused_convolution_4(arg_X, arg_W, out_ptr0):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     KERNEL_H : tl.constexpr = 2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     KERNEL_W : tl.constexpr = 2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     STRIDE_H : tl.constexpr = 2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     STRIDE_W : tl.constexpr = 2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     PADDING_H : tl.constexpr = 0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     PADDING_W : tl.constexpr = 0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     GROUPS : tl.constexpr = 1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     UNROLL : tl.constexpr = False
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     ALLOW_TF32 : tl.constexpr = True
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     BLOCK_M : tl.constexpr = 64
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     BLOCK_N : tl.constexpr = 256
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     BLOCK_K : tl.constexpr = 16
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     X = arg_X
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     W = arg_W
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     # Tensor dimensions
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     BATCH = 2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     IN_C = 4
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     IN_H = 128
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     IN_W = 128
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     OUT_C = 1152
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     OUT_H = 64
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     OUT_W = 64
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     # Strides:
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     stride_xn = 65536
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     stride_xc = 1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     stride_xh = 512
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     stride_xw = 4
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     stride_wc_out = 16
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     stride_wc_in = 1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     stride_wh = 8
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     stride_ww = 4
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     nhw = tl.program_id(0) * BLOCK_M + tl.arange(0, BLOCK_M)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     idx_y_w = nhw % OUT_W
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     nh = nhw // OUT_W
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     idx_y_h = nh % OUT_H
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     idx_n = nh // OUT_H
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     idx_y_c = tl.program_id(1) * BLOCK_N + tl.arange(0, BLOCK_N)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     group = 0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     GROUP_IN_C = IN_C
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     GROUP_OUT_C = OUT_C
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x_base = X + (group * stride_xc * GROUP_IN_C + idx_n * stride_xn)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     w_base = (
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         W + (group * stride_wc_out * GROUP_OUT_C + idx_y_c * stride_wc_out)[None, :]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     acc = tl.zeros((BLOCK_M, BLOCK_N), dtype=tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     # Could be simplified, but slightly slower:
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     # for i in range(KERNEL_H):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     #     for j in range(KERNEL_W):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     #         for k in range(0, GROUP_IN_C, BLOCK_K):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     BLOCK_K_COUNT = (GROUP_IN_C + BLOCK_K - 1) // BLOCK_K
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     for ijk in range(KERNEL_H * KERNEL_W * BLOCK_K_COUNT):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         k = (ijk % BLOCK_K_COUNT) * BLOCK_K
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         ij = ijk // BLOCK_K_COUNT
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         i = ij // KERNEL_W
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         j = ij % KERNEL_W
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         idx_x_h = i - PADDING_H + idx_y_h * STRIDE_H
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         idx_x_w = j - PADDING_W + idx_y_w * STRIDE_W
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         idx_x_c = tl.arange(0, BLOCK_K) + k
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         x_ptrs = x_base + (
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             (idx_x_h * stride_xh)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             + (idx_x_w * stride_xw)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             + (idx_x_c * stride_xc)[None, :]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         mask_x = (
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             (idx_n < BATCH)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             & (idx_x_h >= 0)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             & (idx_x_h < IN_H)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             & (idx_x_w >= 0)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             & (idx_x_w < IN_W)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             & (idx_x_c < GROUP_IN_C)[None, :]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         matrix_x = tl.load(x_ptrs, mask=mask_x, other=0.0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         w_ptrs = w_base + (
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             (idx_x_c * stride_wc_in)[:, None] + (i * stride_wh) + (j * stride_ww)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         mask_w = (idx_x_c[:, None] < GROUP_IN_C) & (idx_y_c[None, :] < GROUP_OUT_C)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         matrix_w = tl.load(w_ptrs, mask=mask_w, other=0.0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         acc += tl.dot(matrix_x, matrix_w, allow_tf32=ALLOW_TF32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     mask = (
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         (idx_n < BATCH)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         & (idx_y_h < OUT_H)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         & (idx_y_w < OUT_W)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         & (idx_y_c < GROUP_OUT_C)[None, :]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     idx_n = idx_n[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     idx_c = idx_y_c[None, :] + group * GROUP_OUT_C
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     idx_h = idx_y_h[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     idx_w = idx_y_w[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     # inductor generates a suffix
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = idx_w + (64*idx_h) + (4096*idx_c) + (4718592*idx_n)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (tl.broadcast_to(idx_c + (1152*idx_w) + (73728*idx_h) + (4718592*idx_n), acc.shape)), acc, mask)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import torch._inductor.kernel.conv
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] meta0 = {'KERNEL_H': 2, 'KERNEL_W': 2, 'STRIDE_H': 2, 'STRIDE_W': 2, 'PADDING_H': 0, 'PADDING_W': 0, 'GROUPS': 1, 'UNROLL': False, 'ALLOW_TF32': True, 'BLOCK_M': 64, 'BLOCK_N': 256, 'BLOCK_K': 16}
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/yr/cyr2uidmv2qdwsxvct57mtyzyd3exxi6oyxcoqocjk6j7r2yfqgu.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add, add_2, mul_4, norm_hidden_states, norm_hidden_states_1], Original ATen: [aten.add, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # add => add
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # add_2 => add_11
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # mul_4 => mul_14
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # norm_hidden_states => add_10, convert_element_type_10, convert_element_type_11, mul_13, rsqrt, sub_1, var_mean
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # norm_hidden_states_1 => add_12
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_mul_native_layer_norm_5 = async_compile.triton('triton_red_fused_add_mul_native_layer_norm_5', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.reduction(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[8192, 2048],
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     reduction_hint=ReductionHint.INNER,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: '*bf16', 7: 'i32', 8: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6, 7, 8), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_red_fused_add_mul_native_layer_norm_5', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 12, 'num_reduction': 2, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_red_fused_add_mul_native_layer_norm_5(in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, in_ptr5, out_ptr2, xnumel, rnumel, XBLOCK : tl.constexpr, RBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 8192
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     rnumel = 1152
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     rbase = tl.arange(0, RBLOCK)[None, :]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x3 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 4096
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7_mean = tl.zeros([XBLOCK, RBLOCK], tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7_m2 = tl.zeros([XBLOCK, RBLOCK], tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7_weight = tl.zeros([XBLOCK, RBLOCK], tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     for roffset in range(0, rnumel, RBLOCK):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rindex = roffset + rbase
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rmask = rindex < rnumel
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         r2 = rindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp0 = tl.load(in_ptr0 + (r2 + (1152*x3)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp1 = tl.load(in_ptr1 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp3 = tl.load(in_ptr2 + (r2 + (1152*x0)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp2 = tmp0 + tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp4 = tmp2 + tmp3
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp5 = tmp4.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp6 = tl.broadcast_to(tmp5, [XBLOCK, RBLOCK])
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp7_mean_next, tmp7_m2_next, tmp7_weight_next = triton_helpers.welford_reduce(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             tmp6, tmp7_mean, tmp7_m2, tmp7_weight, roffset == 0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp7_mean = tl.where(rmask, tmp7_mean_next, tmp7_mean)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp7_m2 = tl.where(rmask, tmp7_m2_next, tmp7_m2)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp7_weight = tl.where(rmask, tmp7_weight_next, tmp7_weight)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7_tmp, tmp8_tmp, tmp9_tmp = triton_helpers.welford(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp7_mean, tmp7_m2, tmp7_weight, 1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7 = tmp7_tmp[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp8 = tmp8_tmp[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9 = tmp9_tmp[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x1 = (xindex // 4096)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     for roffset in range(0, rnumel, RBLOCK):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rindex = roffset + rbase
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rmask = rindex < rnumel
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         r2 = rindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp10 = tl.load(in_ptr0 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp11 = tl.load(in_ptr1 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp13 = tl.load(in_ptr2 + (r2 + (1152*x0)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp24 = tl.load(in_ptr3 + (1152 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp25 = tl.load(in_ptr4 + (1152 + r2 + (6912*x1)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp26 = tl.load(in_ptr5 + (1152 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp32 = tl.load(in_ptr3 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp33 = tl.load(in_ptr4 + (r2 + (6912*x1)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp34 = tl.load(in_ptr5 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp12 = tmp10 + tmp11
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp14 = tmp12 + tmp13
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp15 = tmp14.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp16 = tmp15 - tmp7
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp17 = 1152.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp18 = tmp8 / tmp17
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp19 = 1e-06
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp20 = tmp18 + tmp19
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp21 = libdevice.rsqrt(tmp20)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp22 = tmp16 * tmp21
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp23 = tmp22.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp27 = tmp25 + tmp26
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp28 = tmp24 + tmp27
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp29 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp30 = tmp28 + tmp29
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp31 = tmp23 * tmp30
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp35 = tmp33 + tmp34
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp36 = tmp32 + tmp35
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp37 = tmp31 + tmp36
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tl.store(out_ptr2 + (r2 + (1152*x3)), tmp37, rmask)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/wv/cwvinpaomxydoj2prwmcgwqsfbn2qzxt3zbdlxdoavfadffftn52.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key, query, value], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # key => constant_pad_nd_6
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # query => constant_pad_nd_5
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # value => constant_pad_nd_7
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6 = async_compile.triton('triton_poi_fused_constant_pad_nd_6', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[16777216], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_6', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 1, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_6(in_ptr0, out_ptr0, out_ptr1, out_ptr2, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 16777216
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 2048
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x1 = (xindex // 2048)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = x0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.full([1], 1152, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 < tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tl.load(in_ptr0 + (x0 + (1152*x1)), tmp2, other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x2), tmp3, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr1 + (x2), tmp3, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr2 + (x2), tmp3, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/kl/cklcpz72lzisv4b6xh2s5kk47j7r5qxci2awj5uak4xrzefg6b5t.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_4], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_4 => _scaled_dot_product_cudnn_attention
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7 = async_compile.triton('triton_poi_fused__scaled_dot_product_cudnn_attention_7', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[16777216], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused__scaled_dot_product_cudnn_attention_7', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused__scaled_dot_product_cudnn_attention_7(in_out_ptr0, in_ptr0, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 9437184
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 1152
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = tl.load(in_out_ptr0 + (x2), None).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.load(in_ptr0 + (x0), None, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 + tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(in_out_ptr0 + (x2), tmp2, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/xj/cxjmrv4cb4oezymi343qkmkzbtmvx3ao2wbnvim4lfzjcqk42bgb.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_7], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_7 => constant_pad_nd_8
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8 = async_compile.triton('triton_poi_fused_constant_pad_nd_8', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[16777216], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_8', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 1, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_8(in_ptr0, out_ptr0, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 16777216
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 2048
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x1 = (xindex // 2048)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = x0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.full([1], 1152, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 < tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tl.load(in_ptr0 + ((72*(x1 % 4096)) + (294912*((x0 // 72) % 16)) + (4718592*(x1 // 4096)) + (x0 % 72)), tmp2, other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x2), tmp3, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/iz/cizfa5uobinya2cxdo5uzfvrfn6sqrfbvszojkwxvsjr5lf67qcq.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add, attn_output, hidden_states_10, hidden_states_9], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # add => add
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # attn_output => mul_15
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_10 => add_17
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_9 => div_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_9 = async_compile.triton('triton_poi_fused_add_div_mul_9', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[16777216], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: '*bf16', 7: '*bf16', 8: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6, 7, 8), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_add_div_mul_9', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'num_load': 8, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_add_div_mul_9(in_out_ptr0, in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, in_ptr5, in_ptr6, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 9437184
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 1152
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = (xindex // 4718592)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x4 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x5 = xindex % 4718592
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = tl.load(in_ptr0 + (2304 + x0), None, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.load(in_ptr1 + (2304 + x0 + (6912*x2)), None, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tl.load(in_ptr2 + (2304 + x0), None, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp5 = tl.load(in_out_ptr0 + (x4), None).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp6 = tl.load(in_ptr3 + (x0), None, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp11 = tl.load(in_ptr4 + (x4), None).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp12 = tl.load(in_ptr5 + (x0), None, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp14 = tl.load(in_ptr6 + (x5), None, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tmp1 + tmp2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp4 = tmp0 + tmp3
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7 = tmp5 + tmp6
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp8 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9 = tmp7 * tmp8
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp10 = tmp4 * tmp9
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp13 = tmp11 + tmp12
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp15 = tmp13 + tmp14
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp16 = tmp10 + tmp15
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(in_out_ptr0 + (x4), tmp16, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/gx/cgxt5of5zbfg7n4ls3hv3tgu5eykh3o4bgxsm3qo5ofijskh47uh.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_2], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # query_2 => constant_pad_nd_9
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10 = async_compile.triton('triton_poi_fused_constant_pad_nd_10', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[16777216], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_10', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 1, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_10(in_ptr0, out_ptr0, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 16777216
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 2048
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x1 = (xindex // 2048)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = x0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.full([1], 1152, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 < tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tl.load(in_ptr0 + (x0 + (1152*x1)), tmp2, other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x2), tmp3, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/ym/cym22d2xvzw6lcvhvhh7s2xrku7mzzfhdhyvtnoorogbjt3rfjw6.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_3], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_3 => constant_pad_nd_4
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_11 = async_compile.triton('triton_poi_fused_constant_pad_nd_11', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[2097152], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_11', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_11(in_ptr0, in_ptr1, out_ptr0, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 1228800
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 2048
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x1 = (xindex // 2048)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = x0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.full([1], 1152, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 < tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tl.load(in_ptr0 + (x0 + (1152*x1) + (345600*((x1 % 300) // 300))), tmp2, other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp4 = tl.load(in_ptr1 + (x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp5 = tmp3 + tmp4
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp6 = tmp5.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7 = 0.5
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp8 = tmp6 * tmp7
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9 = tmp6 * tmp6
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp10 = tmp9 * tmp6
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp11 = 0.044715
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp12 = tmp10 * tmp11
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp13 = tmp6 + tmp12
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp14 = 0.7978845608028654
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp15 = tmp13 * tmp14
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp16 = libdevice.tanh(tmp15)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp17 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp18 = tmp16 + tmp17
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19 = tmp8 * tmp18
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp20 = tmp19.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp21 = tl.full(tmp20.shape, 0.0, tmp20.dtype)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp22 = tl.where(tmp2, tmp20, tmp21)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x2), tmp22, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/a6/ca6ajgsmiwk7mesiihdpcjbcgvi7mpvxztghix2rqcki4yfgvilo.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_2, key_6, value_2, value_6], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # key_2 => constant_pad_nd_10
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # key_6 => constant_pad_nd_21
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # value_2 => constant_pad_nd_11
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # value_6 => constant_pad_nd_22
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12 = async_compile.triton('triton_poi_fused_constant_pad_nd_12', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[2097152], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_12', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_12(in_ptr0, in_ptr1, out_ptr0, out_ptr1, out_ptr2, out_ptr3, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 1228800
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 2048
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x1 = (xindex // 2048)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = x0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.full([1], 1152, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 < tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tl.load(in_ptr0 + (x0 + (1152*x1) + (345600*((x1 % 300) // 300))), tmp2, other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp4 = tl.load(in_ptr1 + (x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp5 = tmp3 + tmp4
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp6 = tl.full(tmp5.shape, 0.0, tmp5.dtype)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7 = tl.where(tmp2, tmp5, tmp6)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x2), tmp7, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr1 + (x2), tmp7, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr2 + (x2), tmp7, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr3 + (x2), tmp7, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/gr/cgrbxhvlchnp2nbuyeulry4myfybk7cilzyuhudqywuye235zggz.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_11], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_11 => _scaled_dot_product_efficient_attention
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13 = async_compile.triton('triton_poi_fused__scaled_dot_product_efficient_attention_13', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[1048576], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused__scaled_dot_product_efficient_attention_13', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused__scaled_dot_product_efficient_attention_13(in_out_ptr0, in_ptr0, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 691200
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = xindex < xnumel
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 1152
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = tl.load(in_out_ptr0 + (x2), xmask).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.load(in_ptr0 + (x0), xmask, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 + tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(in_out_ptr0 + (x2), tmp2, xmask)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/vi/cvi62zxotjcncijppau52dxzd6rteio363jy2eac3cyt6qr4vvza.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_11, hidden_states_30], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_11 => constant_pad_nd_12
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_30 => constant_pad_nd_23
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14 = async_compile.triton('triton_poi_fused_constant_pad_nd_14', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[16384], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*i64', 1: '*bf16', 2: '*bf16', 3: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_14', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 1, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_14(in_ptr0, out_ptr0, out_ptr1, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 9728
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = xindex < xnumel
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 304
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = (xindex // 4864)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x3 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = x0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.full([1], 300, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 < tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tl.load(in_ptr0 + (x0 + (300*x2)), tmp2 & xmask, eviction_policy='evict_last', other=0.0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp4 = tmp3.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp5 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp6 = tmp5 - tmp4
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7 = -10000.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp8 = tmp6 * tmp7
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9 = tl.full(tmp8.shape, 0.0, tmp8.dtype)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp10 = tl.where(tmp2, tmp8, tmp9)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x3), tmp10, xmask)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr1 + (x3), tmp10, xmask)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/jn/cjntrr3amnv7csjvqrynnmbehsyli26bzls3getkhv64euuyd5jl.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_6, hidden_states_16, hidden_states_17, mul_6, norm_hidden_states_2, norm_hidden_states_3], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # add_6 => add_24
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_16 => div_2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_17 => add_22
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # mul_6 => mul_17
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # norm_hidden_states_2 => add_23, convert_element_type_12, convert_element_type_13, mul_16, rsqrt_1, sub_2, var_mean_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # norm_hidden_states_3 => add_25
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15 = async_compile.triton('triton_red_fused_add_div_mul_native_layer_norm_15', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.reduction(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[8192, 2048],
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     reduction_hint=ReductionHint.INNER,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: '*bf16', 7: 'i32', 8: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6, 7, 8), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_red_fused_add_div_mul_native_layer_norm_15', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 12, 'num_reduction': 2, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_red_fused_add_div_mul_native_layer_norm_15(in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, in_ptr5, out_ptr2, xnumel, rnumel, XBLOCK : tl.constexpr, RBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 8192
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     rnumel = 1152
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     rbase = tl.arange(0, RBLOCK)[None, :]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9_mean = tl.zeros([XBLOCK, RBLOCK], tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9_m2 = tl.zeros([XBLOCK, RBLOCK], tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9_weight = tl.zeros([XBLOCK, RBLOCK], tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     for roffset in range(0, rnumel, RBLOCK):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rindex = roffset + rbase
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rmask = rindex < rnumel
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         r1 = rindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp0 = tl.load(in_ptr0 + (r1 + (1152*x0)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp1 = tl.load(in_ptr1 + (r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp5 = tl.load(in_ptr2 + (r1 + (1152*x0)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp2 = tmp0 + tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp3 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp4 = tmp2 * tmp3
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp6 = tmp4 + tmp5
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp7 = tmp6.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp8 = tl.broadcast_to(tmp7, [XBLOCK, RBLOCK])
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp9_mean_next, tmp9_m2_next, tmp9_weight_next = triton_helpers.welford_reduce(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             tmp8, tmp9_mean, tmp9_m2, tmp9_weight, roffset == 0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp9_mean = tl.where(rmask, tmp9_mean_next, tmp9_mean)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp9_m2 = tl.where(rmask, tmp9_m2_next, tmp9_m2)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp9_weight = tl.where(rmask, tmp9_weight_next, tmp9_weight)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9_tmp, tmp10_tmp, tmp11_tmp = triton_helpers.welford(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp9_mean, tmp9_m2, tmp9_weight, 1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9 = tmp9_tmp[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp10 = tmp10_tmp[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp11 = tmp11_tmp[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x3 = (xindex // 4096)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     for roffset in range(0, rnumel, RBLOCK):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rindex = roffset + rbase
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rmask = rindex < rnumel
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         r1 = rindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp12 = tl.load(in_ptr0 + (r1 + (1152*x0)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp13 = tl.load(in_ptr1 + (r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp17 = tl.load(in_ptr2 + (r1 + (1152*x0)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp28 = tl.load(in_ptr3 + (4608 + r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp29 = tl.load(in_ptr4 + (4608 + r1 + (6912*x3)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp30 = tl.load(in_ptr5 + (4608 + r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp35 = tl.load(in_ptr3 + (3456 + r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp36 = tl.load(in_ptr4 + (3456 + r1 + (6912*x3)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp37 = tl.load(in_ptr5 + (3456 + r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp14 = tmp12 + tmp13
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp15 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp16 = tmp14 * tmp15
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp18 = tmp16 + tmp17
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp19 = tmp18.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp20 = tmp19 - tmp9
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp21 = 1152.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp22 = tmp10 / tmp21
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp23 = 1e-06
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp24 = tmp22 + tmp23
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp25 = libdevice.rsqrt(tmp24)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp26 = tmp20 * tmp25
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp27 = tmp26.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp31 = tmp29 + tmp30
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp32 = tmp28 + tmp31
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp33 = tmp32 + tmp15
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp34 = tmp27 * tmp33
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp38 = tmp36 + tmp37
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp39 = tmp35 + tmp38
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp40 = tmp34 + tmp39
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tl.store(out_ptr2 + (r1 + (1152*x0)), tmp40, rmask)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/kq/ckqn7d2kqemtlcbocxuabe6qk2zk2b5smq2wimkpq2klicoapdj3.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_21], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_21 => constant_pad_nd_15
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16 = async_compile.triton('triton_poi_fused_constant_pad_nd_16', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[67108864], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_16', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_16(in_ptr0, in_ptr1, out_ptr0, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 41943040
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 5120
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x1 = (xindex // 5120)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = x0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.full([1], 4608, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 < tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tl.load(in_ptr0 + (x0 + (4608*x1) + (18874368*((x1 % 4096) // 4096))), tmp2, other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp4 = tl.load(in_ptr1 + (x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp5 = tmp3 + tmp4
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp6 = tmp5.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7 = 0.5
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp8 = tmp6 * tmp7
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9 = tmp6 * tmp6
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp10 = tmp9 * tmp6
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp11 = 0.044715
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp12 = tmp10 * tmp11
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp13 = tmp6 + tmp12
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp14 = 0.7978845608028654
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp15 = tmp13 * tmp14
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp16 = libdevice.tanh(tmp15)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp17 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp18 = tmp16 + tmp17
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19 = tmp8 * tmp18
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp20 = tmp19.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp21 = tl.full(tmp20.shape, 0.0, tmp20.dtype)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp22 = tl.where(tmp2, tmp20, tmp21)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x2), tmp22, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/ub/cuby5avq3u33nycjmmxevxnqqpmqsqqyi5tm5eh5ywizt5dixnbj.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_10, ff_output, hidden_states_16, hidden_states_17, hidden_states_22, mul_8, norm_hidden_states_4, norm_hidden_states_5], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # add_10 => add_33
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # ff_output => mul_24
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_16 => div_2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_17 => add_22
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_22 => add_30
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # mul_8 => mul_26
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # norm_hidden_states_4 => add_32, convert_element_type_16, convert_element_type_17, mul_25, rsqrt_2, sub_3, var_mean_2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # norm_hidden_states_5 => add_34
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17 = async_compile.triton('triton_red_fused_add_div_mul_native_layer_norm_17', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.reduction(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[8192, 2048],
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     reduction_hint=ReductionHint.INNER,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: '*bf16', 7: '*bf16', 8: '*bf16', 9: '*bf16', 10: 'i32', 11: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_red_fused_add_div_mul_native_layer_norm_17', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'num_load': 15, 'num_reduction': 2, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_red_fused_add_div_mul_native_layer_norm_17(in_out_ptr0, in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, in_ptr5, in_ptr6, in_ptr7, out_ptr2, xnumel, rnumel, XBLOCK : tl.constexpr, RBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 8192
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     rnumel = 1152
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     rbase = tl.arange(0, RBLOCK)[None, :]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x1 = (xindex // 4096)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x3 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19_mean = tl.zeros([XBLOCK, RBLOCK], tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19_m2 = tl.zeros([XBLOCK, RBLOCK], tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19_weight = tl.zeros([XBLOCK, RBLOCK], tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     for roffset in range(0, rnumel, RBLOCK):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rindex = roffset + rbase
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rmask = rindex < rnumel
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         r2 = rindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp0 = tl.load(in_ptr0 + (5760 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp1 = tl.load(in_ptr1 + (5760 + r2 + (6912*x1)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp2 = tl.load(in_ptr2 + (5760 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp5 = tl.load(in_out_ptr0 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp6 = tl.load(in_ptr3 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp9 = tl.load(in_ptr4 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp10 = tl.load(in_ptr5 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp14 = tl.load(in_ptr6 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp3 = tmp1 + tmp2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp4 = tmp0 + tmp3
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp7 = tmp5 + tmp6
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp8 = tmp4 * tmp7
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp11 = tmp9 + tmp10
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp12 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp13 = tmp11 * tmp12
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp15 = tmp13 + tmp14
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp16 = tmp8 + tmp15
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp17 = tmp16.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp18 = tl.broadcast_to(tmp17, [XBLOCK, RBLOCK])
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp19_mean_next, tmp19_m2_next, tmp19_weight_next = triton_helpers.welford_reduce(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             tmp18, tmp19_mean, tmp19_m2, tmp19_weight, roffset == 0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp19_mean = tl.where(rmask, tmp19_mean_next, tmp19_mean)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp19_m2 = tl.where(rmask, tmp19_m2_next, tmp19_m2)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp19_weight = tl.where(rmask, tmp19_weight_next, tmp19_weight)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tl.store(in_out_ptr0 + (r2 + (1152*x3)), tmp16, rmask)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19_tmp, tmp20_tmp, tmp21_tmp = triton_helpers.welford(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp19_mean, tmp19_m2, tmp19_weight, 1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19 = tmp19_tmp[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp20 = tmp20_tmp[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp21 = tmp21_tmp[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     for roffset in range(0, rnumel, RBLOCK):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rindex = roffset + rbase
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rmask = rindex < rnumel
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         r2 = rindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp22 = tl.load(in_out_ptr0 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp32 = tl.load(in_ptr7 + (1152 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp33 = tl.load(in_ptr1 + (1152 + r2 + (6912*x1)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp34 = tl.load(in_ptr2 + (1152 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp40 = tl.load(in_ptr7 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp41 = tl.load(in_ptr1 + (r2 + (6912*x1)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp42 = tl.load(in_ptr2 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp23 = tmp22.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp24 = tmp23 - tmp19
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp25 = 1152.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp26 = tmp20 / tmp25
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp27 = 1e-06
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp28 = tmp26 + tmp27
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp29 = libdevice.rsqrt(tmp28)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp30 = tmp24 * tmp29
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp31 = tmp30.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp35 = tmp33 + tmp34
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp36 = tmp32 + tmp35
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp37 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp38 = tmp36 + tmp37
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp39 = tmp31 * tmp38
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp43 = tmp41 + tmp42
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp44 = tmp40 + tmp43
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp45 = tmp39 + tmp44
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tl.store(out_ptr2 + (r2 + (1152*x3)), tmp45, rmask)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/zl/czlcsoiyra2yl35d6vc2kjmszi72tek4i7g56lw2vawqn2tvlv63.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_1, hidden_states_28, hidden_states_29], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # attn_output_1 => mul_27
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_28 => div_3
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_29 => add_39
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18 = async_compile.triton('triton_poi_fused_add_div_mul_18', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[16777216], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_add_div_mul_18', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'num_load': 6, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_add_div_mul_18(in_out_ptr0, in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 9437184
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 1152
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = (xindex // 4718592)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x4 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = tl.load(in_ptr0 + (2304 + x0), None, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.load(in_ptr1 + (2304 + x0 + (6912*x2)), None, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tl.load(in_ptr2 + (2304 + x0), None, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp5 = tl.load(in_out_ptr0 + (x4), None).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp6 = tl.load(in_ptr3 + (x0), None, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp11 = tl.load(in_ptr4 + (x4), None).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tmp1 + tmp2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp4 = tmp0 + tmp3
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7 = tmp5 + tmp6
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp8 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9 = tmp7 * tmp8
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp10 = tmp4 * tmp9
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp12 = tmp10 + tmp11
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(in_out_ptr0 + (x4), tmp12, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/ks/ckskwpln7okqblhb2s5avhs6d6x67kskjaaoywrtq7pc7qyenaes.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [ff_output_27, hidden_states_529, hidden_states_530, hidden_states_535, hidden_states_536], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # ff_output_27 => mul_348
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_529 => div_56
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_530 => add_616
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_535 => add_624
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_536 => convert_element_type_178, var_mean_56
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_19 = async_compile.triton('triton_red_fused_add_div_mul_native_layer_norm_19', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.reduction(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[8192, 2048],
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     reduction_hint=ReductionHint.INNER,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: '*bf16', 7: '*bf16', 8: '*fp32', 9: '*fp32', 10: 'i32', 11: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_red_fused_add_div_mul_native_layer_norm_19', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'num_load': 8, 'num_reduction': 2, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_red_fused_add_div_mul_native_layer_norm_19(in_out_ptr0, in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, in_ptr5, in_ptr6, out_ptr0, out_ptr1, xnumel, rnumel, XBLOCK : tl.constexpr, RBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 8192
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     rnumel = 1152
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     rbase = tl.arange(0, RBLOCK)[None, :]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x1 = (xindex // 4096)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x3 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19_mean = tl.zeros([XBLOCK, RBLOCK], tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19_m2 = tl.zeros([XBLOCK, RBLOCK], tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19_weight = tl.zeros([XBLOCK, RBLOCK], tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     for roffset in range(0, rnumel, RBLOCK):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rindex = roffset + rbase
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         rmask = rindex < rnumel
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         r2 = rindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp0 = tl.load(in_ptr0 + (5760 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp1 = tl.load(in_ptr1 + (5760 + r2 + (6912*x1)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp2 = tl.load(in_ptr2 + (5760 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp5 = tl.load(in_out_ptr0 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp6 = tl.load(in_ptr3 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp9 = tl.load(in_ptr4 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp10 = tl.load(in_ptr5 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp14 = tl.load(in_ptr6 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp3 = tmp1 + tmp2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp4 = tmp0 + tmp3
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp7 = tmp5 + tmp6
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp8 = tmp4 * tmp7
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp11 = tmp9 + tmp10
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp12 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp13 = tmp11 * tmp12
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp15 = tmp13 + tmp14
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp16 = tmp8 + tmp15
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp17 = tmp16.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp18 = tl.broadcast_to(tmp17, [XBLOCK, RBLOCK])
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp19_mean_next, tmp19_m2_next, tmp19_weight_next = triton_helpers.welford_reduce(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]             tmp18, tmp19_mean, tmp19_m2, tmp19_weight, roffset == 0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp19_mean = tl.where(rmask, tmp19_mean_next, tmp19_mean)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp19_m2 = tl.where(rmask, tmp19_m2_next, tmp19_m2)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp19_weight = tl.where(rmask, tmp19_weight_next, tmp19_weight)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tl.store(in_out_ptr0 + (r2 + (1152*x3)), tmp16, rmask)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19_tmp, tmp20_tmp, tmp21_tmp = triton_helpers.welford(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         tmp19_mean, tmp19_m2, tmp19_weight, 1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19 = tmp19_tmp[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp20 = tmp20_tmp[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp21 = tmp21_tmp[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x3), tmp19, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr1 + (x3), tmp20, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/yl/cylfrobtjz27y6zkonkaeawdjman53mf5gful2kwrjzkngczzxz3.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_538], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_538 => constant_pad_nd_313
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_20 = async_compile.triton('triton_poi_fused_constant_pad_nd_20', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[16777216], 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*fp32', 2: '*fp32', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: '*bf16', 7: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6, 7), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_20', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 7, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_20(in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, in_ptr5, out_ptr0, xnumel, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 16777216
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x0 = xindex % 2048
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x1 = (xindex // 2048)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = x0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.full([1], 1152, tl.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 < tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp3 = tl.load(in_ptr0 + (x0 + (1152*x1)), tmp2, other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp4 = tmp3.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp5 = tl.load(in_ptr1 + (x1), tmp2, eviction_policy='evict_last', other=0.0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp6 = tmp4 - tmp5
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp7 = tl.load(in_ptr2 + (x1), tmp2, eviction_policy='evict_last', other=0.0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp8 = 1152.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp9 = tmp7 / tmp8
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp10 = 1e-06
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp11 = tmp9 + tmp10
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp12 = libdevice.rsqrt(tmp11)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp13 = tmp6 * tmp12
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp14 = tmp13.to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp15 = tl.load(in_ptr3 + (1152 + x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp16 = tl.load(in_ptr4 + (x0 + (1152*(x1 // 4096))), tmp2, other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp17 = tl.load(in_ptr5 + (x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp18 = tmp16 + tmp17
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp19 = tmp15 + tmp18
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp20 = 1.0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp21 = tmp19 + tmp20
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp22 = tmp14 * tmp21
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp23 = tl.load(in_ptr3 + (x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp24 = tmp23 + tmp18
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp25 = tmp22 + tmp24
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp26 = tl.full(tmp25.shape, 0.0, tmp25.dtype)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp27 = tl.where(tmp2, tmp25, tmp26)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x2), tmp27, None)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/fk/cfksz3ijwl65jz33sewtbiwkc5sbwm7cr5juqhol74ejxyb4tijh.py
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [output], Original ATen: [aten.clone]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # output => clone_140
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_clone_21 = async_compile.triton('triton_poi_fused_clone_21', '''
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise(
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     size_hints=[16, 16384], tile_hint=TileHint.DEFAULT,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     filename=__file__,
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: 'i32', 4: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4), equal_to_1=())]},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_clone_21', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True},
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     min_elem_per_thread=0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_clone_21(in_ptr0, in_ptr1, out_ptr0, ynumel, xnumel, YBLOCK : tl.constexpr, XBLOCK : tl.constexpr):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     ynumel = 16
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xnumel = 16384
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     yoffset = tl.program_id(1) * YBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     yindex = yoffset + tl.arange(0, YBLOCK)[None, :]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     ymask = yindex < ynumel
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xoffset = tl.program_id(0) * XBLOCK
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xindex = xoffset + tl.arange(0, XBLOCK)[:, None]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     xmask = tl.full([XBLOCK, YBLOCK], True, tl.int1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x2 = xindex % 2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x3 = (xindex // 2) % 64
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x4 = (xindex // 128) % 2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x5 = (xindex // 256)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     y0 = yindex % 8
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     y1 = (yindex // 8)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     x7 = xindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     y6 = yindex
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp0 = tl.load(in_ptr0 + (y0 + (8*x2) + (16*x4) + (32*x3) + (2048*x5) + (131072*y1) + (131072*((x3 + (64*x5)) // 4096))), ymask, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp1 = tl.load(in_ptr1 + (y0 + (8*x2) + (16*x4)), ymask, eviction_policy='evict_last').to(tl.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tmp2 = tmp0 + tmp1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     tl.store(out_ptr0 + (x7 + (16384*y6)), tmp2, ymask)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda')
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] async_compile.wait(globals())
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del async_compile
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def call(args):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg0_1, arg1_1, arg2_1, arg3_1, arg4_1, arg5_1, arg6_1, arg7_1, arg8_1, arg9_1, arg10_1, arg11_1, arg12_1, arg13_1, arg14_1, arg15_1, arg16_1, arg17_1, arg18_1, arg19_1, arg20_1, arg21_1, arg22_1, arg23_1, arg24_1, arg25_1, arg26_1, arg27_1, arg28_1, arg29_1, arg30_1, arg31_1, arg32_1, arg33_1, arg34_1, arg35_1, arg36_1, arg37_1, arg38_1, arg39_1, arg40_1, arg41_1, arg42_1, arg43_1, arg44_1, arg45_1, arg46_1, arg47_1, arg48_1, arg49_1, arg50_1, arg51_1, arg52_1, arg53_1, arg54_1, arg55_1, arg56_1, arg57_1, arg58_1, arg59_1, arg60_1, arg61_1, arg62_1, arg63_1, arg64_1, arg65_1, arg66_1, arg67_1, arg68_1, arg69_1, arg70_1, arg71_1, arg72_1, arg73_1, arg74_1, arg75_1, arg76_1, arg77_1, arg78_1, arg79_1, arg80_1, arg81_1, arg82_1, arg83_1, arg84_1, arg85_1, arg86_1, arg87_1, arg88_1, arg89_1, arg90_1, arg91_1, arg92_1, arg93_1, arg94_1, arg95_1, arg96_1, arg97_1, arg98_1, arg99_1, arg100_1, arg101_1, arg102_1, arg103_1, arg104_1, arg105_1, arg106_1, arg107_1, arg108_1, arg109_1, arg110_1, arg111_1, arg112_1, arg113_1, arg114_1, arg115_1, arg116_1, arg117_1, arg118_1, arg119_1, arg120_1, arg121_1, arg122_1, arg123_1, arg124_1, arg125_1, arg126_1, arg127_1, arg128_1, arg129_1, arg130_1, arg131_1, arg132_1, arg133_1, arg134_1, arg135_1, arg136_1, arg137_1, arg138_1, arg139_1, arg140_1, arg141_1, arg142_1, arg143_1, arg144_1, arg145_1, arg146_1, arg147_1, arg148_1, arg149_1, arg150_1, arg151_1, arg152_1, arg153_1, arg154_1, arg155_1, arg156_1, arg157_1, arg158_1, arg159_1, arg160_1, arg161_1, arg162_1, arg163_1, arg164_1, arg165_1, arg166_1, arg167_1, arg168_1, arg169_1, arg170_1, arg171_1, arg172_1, arg173_1, arg174_1, arg175_1, arg176_1, arg177_1, arg178_1, arg179_1, arg180_1, arg181_1, arg182_1, arg183_1, arg184_1, arg185_1, arg186_1, arg187_1, arg188_1, arg189_1, arg190_1, arg191_1, arg192_1, arg193_1, arg194_1, arg195_1, arg196_1, arg197_1, arg198_1, arg199_1, arg200_1, arg201_1, arg202_1, arg203_1, arg204_1, arg205_1, arg206_1, arg207_1, arg208_1, arg209_1, arg210_1, arg211_1, arg212_1, arg213_1, arg214_1, arg215_1, arg216_1, arg217_1, arg218_1, arg219_1, arg220_1, arg221_1, arg222_1, arg223_1, arg224_1, arg225_1, arg226_1, arg227_1, arg228_1, arg229_1, arg230_1, arg231_1, arg232_1, arg233_1, arg234_1, arg235_1, arg236_1, arg237_1, arg238_1, arg239_1, arg240_1, arg241_1, arg242_1, arg243_1, arg244_1, arg245_1, arg246_1, arg247_1, arg248_1, arg249_1, arg250_1, arg251_1, arg252_1, arg253_1, arg254_1, arg255_1, arg256_1, arg257_1, arg258_1, arg259_1, arg260_1, arg261_1, arg262_1, arg263_1, arg264_1, arg265_1, arg266_1, arg267_1, arg268_1, arg269_1, arg270_1, arg271_1, arg272_1, arg273_1, arg274_1, arg275_1, arg276_1, arg277_1, arg278_1, arg279_1, arg280_1, arg281_1, arg282_1, arg283_1, arg284_1, arg285_1, arg286_1, arg287_1, arg288_1, arg289_1, arg290_1, arg291_1, arg292_1, arg293_1, arg294_1, arg295_1, arg296_1, arg297_1, arg298_1, arg299_1, arg300_1, arg301_1, arg302_1, arg303_1, arg304_1, arg305_1, arg306_1, arg307_1, arg308_1, arg309_1, arg310_1, arg311_1, arg312_1, arg313_1, arg314_1, arg315_1, arg316_1, arg317_1, arg318_1, arg319_1, arg320_1, arg321_1, arg322_1, arg323_1, arg324_1, arg325_1, arg326_1, arg327_1, arg328_1, arg329_1, arg330_1, arg331_1, arg332_1, arg333_1, arg334_1, arg335_1, arg336_1, arg337_1, arg338_1, arg339_1, arg340_1, arg341_1, arg342_1, arg343_1, arg344_1, arg345_1, arg346_1, arg347_1, arg348_1, arg349_1, arg350_1, arg351_1, arg352_1, arg353_1, arg354_1, arg355_1, arg356_1, arg357_1, arg358_1, arg359_1, arg360_1, arg361_1, arg362_1, arg363_1, arg364_1, arg365_1, arg366_1, arg367_1, arg368_1, arg369_1, arg370_1, arg371_1, arg372_1, arg373_1, arg374_1, arg375_1, arg376_1, arg377_1, arg378_1, arg379_1, arg380_1, arg381_1, arg382_1, arg383_1, arg384_1, arg385_1, arg386_1, arg387_1, arg388_1, arg389_1, arg390_1, arg391_1, arg392_1, arg393_1, arg394_1, arg395_1, arg396_1, arg397_1, arg398_1, arg399_1, arg400_1, arg401_1, arg402_1, arg403_1, arg404_1, arg405_1, arg406_1, arg407_1, arg408_1, arg409_1, arg410_1, arg411_1, arg412_1, arg413_1, arg414_1, arg415_1, arg416_1, arg417_1, arg418_1, arg419_1, arg420_1, arg421_1, arg422_1, arg423_1, arg424_1, arg425_1, arg426_1, arg427_1, arg428_1, arg429_1, arg430_1, arg431_1, arg432_1, arg433_1, arg434_1, arg435_1, arg436_1, arg437_1, arg438_1, arg439_1, arg440_1, arg441_1, arg442_1, arg443_1, arg444_1, arg445_1, arg446_1, arg447_1, arg448_1, arg449_1, arg450_1, arg451_1, arg452_1, arg453_1, arg454_1, arg455_1, arg456_1, arg457_1, arg458_1, arg459_1, arg460_1, arg461_1, arg462_1, arg463_1, arg464_1, arg465_1, arg466_1, arg467_1, arg468_1, arg469_1, arg470_1, arg471_1, arg472_1, arg473_1, arg474_1, arg475_1, arg476_1, arg477_1, arg478_1, arg479_1, arg480_1, arg481_1, arg482_1, arg483_1, arg484_1, arg485_1, arg486_1, arg487_1, arg488_1, arg489_1, arg490_1, arg491_1, arg492_1, arg493_1, arg494_1, arg495_1, arg496_1, arg497_1, arg498_1, arg499_1, arg500_1, arg501_1, arg502_1, arg503_1, arg504_1, arg505_1, arg506_1, arg507_1, arg508_1, arg509_1, arg510_1, arg511_1, arg512_1, arg513_1, arg514_1, arg515_1, arg516_1, arg517_1, arg518_1, arg519_1, arg520_1, arg521_1, arg522_1, arg523_1, arg524_1, arg525_1, arg526_1, arg527_1, arg528_1, arg529_1, arg530_1, arg531_1, arg532_1, arg533_1, arg534_1, arg535_1, arg536_1, arg537_1, arg538_1, arg539_1, arg540_1, arg541_1, arg542_1, arg543_1, arg544_1, arg545_1, arg546_1, arg547_1, arg548_1, arg549_1, arg550_1, arg551_1, arg552_1, arg553_1, arg554_1, arg555_1, arg556_1, arg557_1, arg558_1, arg559_1, arg560_1, arg561_1, arg562_1, arg563_1, arg564_1, arg565_1, arg566_1, arg567_1, arg568_1, arg569_1, arg570_1, arg571_1, arg572_1, arg573_1, arg574_1, arg575_1, arg576_1, arg577_1, arg578_1, arg579_1, arg580_1, arg581_1, arg582_1, arg583_1, arg584_1, arg585_1, arg586_1, arg587_1, arg588_1, arg589_1, arg590_1, arg591_1, arg592_1, arg593_1, arg594_1, arg595_1, arg596_1, arg597_1, arg598_1, arg599_1, arg600_1, arg601_1, arg602_1, arg603_1, arg604_1, arg605_1, arg606_1, arg607_1, arg608_1, arg609_1, arg610_1, arg611_1, arg612_1, arg613_1, arg614_1, arg615_1, arg616_1, arg617_1, arg618_1, arg619_1, arg620_1, arg621_1, arg622_1, arg623_1, arg624_1, arg625_1, arg626_1, arg627_1, arg628_1, arg629_1, arg630_1, arg631_1, arg632_1, arg633_1, arg634_1, arg635_1, arg636_1, arg637_1, arg638_1, arg639_1, arg640_1, arg641_1, arg642_1, arg643_1, arg644_1, arg645_1, arg646_1, arg647_1, arg648_1, arg649_1, arg650_1, arg651_1, arg652_1, arg653_1, arg654_1, arg655_1, arg656_1, arg657_1, arg658_1, arg659_1, arg660_1, arg661_1, arg662_1, arg663_1, arg664_1, arg665_1, arg666_1, arg667_1, arg668_1, arg669_1, arg670_1, arg671_1, arg672_1, arg673_1, arg674_1, arg675_1, arg676_1, arg677_1, arg678_1, arg679_1, arg680_1, arg681_1, arg682_1, arg683_1, arg684_1, arg685_1, arg686_1, arg687_1, arg688_1, arg689_1, arg690_1, arg691_1, arg692_1, arg693_1, arg694_1, arg695_1, arg696_1, arg697_1, arg698_1, arg699_1, arg700_1, arg701_1, arg702_1, arg703_1, arg704_1, arg705_1, arg706_1, arg707_1, arg708_1, arg709_1, arg710_1, arg711_1, arg712_1, arg713_1, arg714_1, arg715_1, arg716_1, arg717_1, arg718_1, arg719_1, arg720_1, arg721_1, arg722_1, arg723_1, arg724_1, arg725_1, arg726_1, arg727_1, arg728_1, arg729_1, arg730_1, arg731_1, arg732_1, arg733_1, arg734_1, arg735_1, arg736_1, arg737_1, arg738_1, arg739_1, arg740_1, arg741_1, arg742_1, arg743_1, arg744_1, arg745_1, arg746_1, arg747_1, arg748_1, arg749_1, arg750_1, arg751_1, arg752_1, arg753_1, arg754_1, arg755_1, arg756_1, arg757_1, arg758_1, arg759_1, arg760_1, arg761_1, arg762_1, arg763_1, arg764_1, arg765_1, arg766_1, arg767_1, arg768_1, arg769_1, arg770_1, arg771_1, arg772_1, arg773_1, arg774_1, arg775_1, arg776_1, arg777_1, arg778_1, arg779_1, arg780_1, arg781_1, arg782_1, arg783_1, arg784_1, arg785_1, arg786_1, arg787_1, arg788_1, arg789_1, arg790_1, arg791_1, arg792_1, arg793_1, arg794_1, arg795_1, arg796_1, arg797_1, arg798_1, arg799_1, arg800_1, arg801_1, arg802_1, arg803_1, arg804_1, arg805_1, arg806_1, arg807_1, arg808_1, arg809_1, arg810_1, arg811_1, arg812_1, arg813_1, arg814_1, arg815_1, arg816_1, arg817_1, arg818_1, arg819_1, arg820_1, arg821_1, arg822_1, arg823_1, arg824_1, arg825_1, arg826_1, arg827_1, arg828_1, arg829_1, arg830_1, arg831_1, arg832_1, arg833_1, arg834_1, arg835_1, arg836_1, arg837_1, arg838_1, arg839_1, arg840_1, arg841_1, arg842_1, arg843_1, arg844_1, arg845_1, arg846_1, arg847_1, arg848_1, arg849_1, arg850_1, arg851_1, arg852_1, arg853_1, arg854_1, arg855_1, arg856_1, arg857_1, arg858_1, arg859_1, arg860_1, arg861_1, arg862_1, arg863_1, arg864_1, arg865_1, arg866_1, arg867_1, arg868_1, arg869_1, arg870_1, arg871_1, arg872_1, arg873_1, arg874_1, arg875_1, arg876_1, arg877_1, arg878_1, arg879_1, arg880_1, arg881_1, arg882_1, arg883_1, arg884_1, arg885_1, arg886_1, arg887_1, arg888_1, arg889_1, arg890_1, arg891_1, arg892_1, arg893_1 = args
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     args.clear()
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg0_1, (2, 300), (300, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg1_1, (2, 4, 128, 128), (65536, 16384, 128, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg2_1, (1152, 4, 2, 2), (16, 1, 8, 4))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg3_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg4_1, (1, 4096, 1152), (4718592, 1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg5_1, (2, ), (0, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg6_1, (144, 8, 32, 4), (1024, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg7_1, (8, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg8_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg9_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg10_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg11_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg12_1, (864, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg13_1, (16, 6912, 2), (13824, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg14_1, (6912, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg15_1, (144, 32, 32, 4), (4096, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg16_1, (32, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg17_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg18_1, (2, 300, 4096), (1228800, 4096, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg19_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg20_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg21_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg22_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg23_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg24_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg25_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg26_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg27_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg28_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg29_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg30_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg31_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg32_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg33_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg34_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg35_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg36_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg37_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg38_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg39_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg40_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg41_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg42_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg43_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg44_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg45_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg46_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg47_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg48_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg49_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg50_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg51_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg52_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg53_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg54_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg55_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg56_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg57_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg58_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg59_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg60_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg61_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg62_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg63_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg64_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg65_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg66_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg67_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg68_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg69_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg70_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg71_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg72_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg73_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg74_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg75_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg76_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg77_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg78_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg79_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg80_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg81_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg82_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg83_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg84_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg85_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg86_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg87_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg88_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg89_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg90_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg91_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg92_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg93_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg94_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg95_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg96_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg97_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg98_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg99_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg100_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg101_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg102_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg103_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg104_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg105_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg106_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg107_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg108_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg109_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg110_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg111_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg112_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg113_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg114_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg115_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg116_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg117_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg118_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg119_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg120_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg121_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg122_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg123_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg124_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg125_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg126_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg127_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg128_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg129_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg130_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg131_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg132_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg133_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg134_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg135_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg136_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg137_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg138_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg139_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg140_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg141_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg142_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg143_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg144_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg145_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg146_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg147_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg148_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg149_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg150_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg151_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg152_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg153_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg154_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg155_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg156_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg157_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg158_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg159_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg160_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg161_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg162_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg163_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg164_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg165_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg166_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg167_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg168_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg169_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg170_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg171_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg172_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg173_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg174_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg175_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg176_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg177_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg178_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg179_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg180_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg181_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg182_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg183_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg184_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg185_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg186_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg187_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg188_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg189_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg190_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg191_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg192_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg193_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg194_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg195_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg196_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg197_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg198_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg199_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg200_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg201_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg202_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg203_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg204_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg205_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg206_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg207_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg208_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg209_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg210_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg211_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg212_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg213_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg214_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg215_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg216_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg217_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg218_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg219_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg220_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg221_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg222_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg223_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg224_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg225_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg226_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg227_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg228_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg229_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg230_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg231_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg232_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg233_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg234_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg235_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg236_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg237_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg238_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg239_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg240_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg241_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg242_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg243_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg244_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg245_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg246_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg247_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg248_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg249_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg250_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg251_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg252_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg253_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg254_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg255_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg256_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg257_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg258_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg259_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg260_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg261_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg262_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg263_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg264_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg265_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg266_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg267_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg268_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg269_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg270_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg271_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg272_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg273_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg274_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg275_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg276_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg277_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg278_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg279_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg280_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg281_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg282_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg283_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg284_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg285_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg286_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg287_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg288_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg289_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg290_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg291_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg292_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg293_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg294_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg295_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg296_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg297_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg298_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg299_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg300_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg301_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg302_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg303_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg304_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg305_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg306_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg307_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg308_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg309_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg310_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg311_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg312_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg313_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg314_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg315_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg316_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg317_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg318_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg319_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg320_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg321_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg322_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg323_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg324_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg325_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg326_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg327_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg328_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg329_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg330_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg331_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg332_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg333_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg334_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg335_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg336_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg337_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg338_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg339_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg340_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg341_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg342_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg343_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg344_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg345_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg346_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg347_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg348_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg349_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg350_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg351_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg352_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg353_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg354_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg355_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg356_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg357_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg358_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg359_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg360_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg361_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg362_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg363_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg364_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg365_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg366_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg367_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg368_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg369_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg370_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg371_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg372_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg373_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg374_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg375_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg376_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg377_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg378_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg379_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg380_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg381_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg382_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg383_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg384_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg385_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg386_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg387_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg388_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg389_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg390_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg391_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg392_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg393_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg394_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg395_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg396_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg397_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg398_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg399_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg400_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg401_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg402_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg403_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg404_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg405_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg406_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg407_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg408_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg409_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg410_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg411_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg412_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg413_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg414_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg415_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg416_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg417_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg418_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg419_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg420_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg421_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg422_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg423_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg424_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg425_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg426_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg427_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg428_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg429_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg430_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg431_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg432_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg433_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg434_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg435_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg436_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg437_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg438_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg439_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg440_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg441_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg442_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg443_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg444_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg445_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg446_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg447_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg448_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg449_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg450_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg451_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg452_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg453_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg454_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg455_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg456_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg457_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg458_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg459_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg460_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg461_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg462_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg463_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg464_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg465_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg466_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg467_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg468_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg469_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg470_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg471_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg472_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg473_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg474_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg475_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg476_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg477_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg478_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg479_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg480_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg481_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg482_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg483_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg484_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg485_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg486_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg487_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg488_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg489_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg490_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg491_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg492_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg493_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg494_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg495_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg496_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg497_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg498_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg499_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg500_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg501_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg502_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg503_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg504_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg505_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg506_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg507_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg508_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg509_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg510_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg511_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg512_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg513_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg514_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg515_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg516_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg517_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg518_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg519_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg520_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg521_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg522_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg523_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg524_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg525_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg526_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg527_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg528_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg529_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg530_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg531_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg532_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg533_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg534_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg535_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg536_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg537_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg538_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg539_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg540_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg541_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg542_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg543_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg544_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg545_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg546_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg547_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg548_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg549_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg550_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg551_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg552_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg553_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg554_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg555_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg556_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg557_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg558_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg559_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg560_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg561_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg562_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg563_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg564_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg565_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg566_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg567_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg568_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg569_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg570_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg571_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg572_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg573_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg574_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg575_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg576_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg577_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg578_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg579_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg580_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg581_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg582_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg583_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg584_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg585_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg586_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg587_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg588_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg589_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg590_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg591_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg592_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg593_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg594_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg595_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg596_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg597_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg598_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg599_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg600_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg601_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg602_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg603_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg604_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg605_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg606_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg607_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg608_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg609_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg610_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg611_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg612_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg613_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg614_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg615_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg616_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg617_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg618_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg619_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg620_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg621_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg622_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg623_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg624_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg625_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg626_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg627_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg628_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg629_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg630_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg631_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg632_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg633_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg634_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg635_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg636_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg637_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg638_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg639_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg640_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg641_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg642_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg643_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg644_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg645_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg646_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg647_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg648_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg649_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg650_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg651_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg652_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg653_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg654_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg655_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg656_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg657_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg658_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg659_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg660_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg661_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg662_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg663_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg664_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg665_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg666_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg667_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg668_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg669_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg670_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg671_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg672_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg673_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg674_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg675_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg676_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg677_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg678_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg679_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg680_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg681_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg682_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg683_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg684_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg685_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg686_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg687_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg688_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg689_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg690_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg691_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg692_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg693_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg694_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg695_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg696_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg697_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg698_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg699_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg700_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg701_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg702_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg703_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg704_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg705_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg706_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg707_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg708_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg709_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg710_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg711_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg712_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg713_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg714_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg715_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg716_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg717_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg718_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg719_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg720_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg721_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg722_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg723_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg724_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg725_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg726_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg727_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg728_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg729_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg730_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg731_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg732_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg733_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg734_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg735_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg736_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg737_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg738_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg739_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg740_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg741_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg742_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg743_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg744_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg745_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg746_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg747_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg748_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg749_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg750_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg751_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg752_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg753_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg754_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg755_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg756_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg757_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg758_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg759_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg760_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg761_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg762_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg763_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg764_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg765_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg766_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg767_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg768_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg769_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg770_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg771_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg772_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg773_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg774_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg775_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg776_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg777_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg778_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg779_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg780_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg781_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg782_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg783_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg784_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg785_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg786_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg787_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg788_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg789_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg790_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg791_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg792_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg793_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg794_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg795_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg796_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg797_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg798_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg799_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg800_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg801_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg802_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg803_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg804_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg805_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg806_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg807_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg808_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg809_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg810_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg811_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg812_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg813_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg814_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg815_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg816_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg817_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg818_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg819_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg820_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg821_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg822_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg823_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg824_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg825_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg826_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg827_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg828_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg829_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg830_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg831_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg832_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg833_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg834_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg835_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg836_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg837_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg838_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg839_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg840_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg841_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg842_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg843_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg844_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg845_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg846_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg847_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg848_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg849_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg850_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg851_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg852_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg853_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg854_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg855_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg856_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg857_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg858_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg859_1, (6, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg860_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg861_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg862_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg863_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg864_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg865_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg866_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg867_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg868_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg869_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg870_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg871_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg872_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg873_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg874_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg875_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg876_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg877_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg878_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg879_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg880_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg881_1, (144, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg882_1, (16, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg883_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg884_1, (576, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg885_1, (16, 4608, 2), (9216, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg886_1, (4608, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg887_1, (144, 40, 32, 4), (5120, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg888_1, (40, 1152, 2), (2304, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg889_1, (1152, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg890_1, (2, 1152), (1152, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg891_1, (4, 16, 32, 4), (2048, 128, 4, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg892_1, (16, 32, 2), (64, 2, 1))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     assert_size_stride(arg893_1, (32, ), (1, ))
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     with torch.cuda._DeviceGuard(0):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         torch.cuda.set_device(0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf0 = empty_strided_cuda((2, 256), (256, 1), torch.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [emb_3], Original ATen: [aten.cat]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         stream0 = get_raw_stream(0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_cat_0.run(arg5_1, buf0, 512, grid=grid(512), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg5_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1 = empty_strided_cuda((2, 1024), (1024, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [emb_4, sample, to_2], Original ATen: [aten._to_copy, aten.cat, aten.constant_pad_nd, aten.view]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__to_copy_cat_constant_pad_nd_view_1.run(buf0, buf1, 2048, grid=grid(2048), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf0
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [emb_4, sample, to_2], Original ATen: [aten._to_copy, aten._weight_int4pack_mm, aten.cat, aten.constant_pad_nd, aten.view]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf2 = torch.ops.aten._weight_int4pack_mm.default(buf1, arg6_1, 128, arg7_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg6_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg7_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf3 = buf2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf2
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf4 = empty_strided_cuda((2, 2048), (2048, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [sample, sample_1, sample_2], Original ATen: [aten.add, aten.constant_pad_nd, aten.silu, aten.view]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_constant_pad_nd_silu_view_2.run(buf3, arg8_1, buf4, 4096, grid=grid(4096), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg8_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf3
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [sample, sample_1, sample_2], Original ATen: [aten._weight_int4pack_mm, aten.add, aten.constant_pad_nd, aten.silu, aten.view]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf5 = torch.ops.aten._weight_int4pack_mm.default(buf4, arg9_1, 128, arg10_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg10_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg9_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf6 = buf5
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf5
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf7 = buf4; del buf4  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [sample_2, silu_1, timestep], Original ATen: [aten.add, aten.constant_pad_nd, aten.silu, aten.view]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_constant_pad_nd_silu_view_2.run(buf6, arg11_1, buf7, 4096, grid=grid(4096), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [sample_2, silu_1, timestep], Original ATen: [aten._weight_int4pack_mm, aten.add, aten.constant_pad_nd, aten.silu, aten.view]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf8 = torch.ops.aten._weight_int4pack_mm.default(buf7, arg12_1, 128, arg13_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg12_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg13_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf7
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf9 = buf8
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf8
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf10 = empty_strided_cuda((2, 4, 128, 128), (65536, 1, 512, 4), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [latent], Original ATen: [aten.convolution]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_convolution_3.run(arg1_1, buf10, 8, 16384, grid=grid(8, 16384), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg1_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf11 = empty_strided_cuda((2, 1152, 64, 64), (4718592, 1, 73728, 1152), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [latent], Original ATen: [aten.convolution]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_tem_fused_convolution_4.run(buf10, arg2_1, buf11, grid=torch._inductor.kernel.conv.conv2d_grid(2, 1152, 64, 64, meta0), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg2_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf10
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf15 = empty_strided_cuda((2, 4096, 1152), (4718592, 1152, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add, add_2, mul_4, norm_hidden_states, norm_hidden_states_1], Original ATen: [aten.add, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_mul_native_layer_norm_5.run(buf11, arg3_1, arg4_1, arg22_1, buf9, arg14_1, buf15, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf16 = empty_strided_cuda((8192, 2048), (2048, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf19 = empty_strided_cuda((8192, 2048), (2048, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf22 = empty_strided_cuda((8192, 2048), (2048, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key, query, value], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf15, buf16, buf19, buf22, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf15
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf17 = torch.ops.aten._weight_int4pack_mm.default(buf16, arg23_1, 128, arg24_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg23_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg24_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf18 = buf17
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf17
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf20 = torch.ops.aten._weight_int4pack_mm.default(buf19, arg26_1, 128, arg27_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg26_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg27_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf21 = buf20
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf20
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf23 = torch.ops.aten._weight_int4pack_mm.default(buf22, arg29_1, 128, arg30_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg29_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg30_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf24 = buf23
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf23
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf25 = reinterpret_tensor(buf18, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf18  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_4], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf25, arg25_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg25_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf26 = reinterpret_tensor(buf21, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf21  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_4], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf26, arg28_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg28_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf27 = reinterpret_tensor(buf24, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf24  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_4], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf27, arg31_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg31_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_4], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf28 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf25, buf26, buf27, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf25
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf26
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf27
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf29 = buf28[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf28
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf33 = buf22; del buf22  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_7], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf29, buf33, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf29
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_7], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf34 = torch.ops.aten._weight_int4pack_mm.default(buf33, arg32_1, 128, arg33_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg32_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg33_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf35 = buf34
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf34
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf36 = reinterpret_tensor(buf35, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf35  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add, attn_output, hidden_states_10, hidden_states_9], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_9.run(buf36, arg22_1, buf9, arg14_1, arg34_1, buf11, arg3_1, arg4_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg34_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg3_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg4_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf11
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf37 = buf33; del buf33  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_2], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf36, buf37, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_2], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf38 = torch.ops.aten._weight_int4pack_mm.default(buf37, arg35_1, 128, arg36_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg35_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg36_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf39 = buf38
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf38
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_1], Original ATen: [aten._weight_int4pack_mm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf40 = torch.ops.aten._weight_int4pack_mm.default(reinterpret_tensor(arg18_1, (600, 4096), (4096, 1), 0), arg15_1, 128, arg16_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg15_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg16_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg18_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf41 = buf40
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf40
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf42 = empty_strided_cuda((600, 2048), (2048, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_3], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_11.run(buf41, arg17_1, buf42, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg17_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf41
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_3], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf43 = torch.ops.aten._weight_int4pack_mm.default(buf42, arg19_1, 128, arg20_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg19_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg20_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf44 = buf43
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf43
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf45 = buf42; del buf42  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf48 = empty_strided_cuda((600, 2048), (2048, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf102 = empty_strided_cuda((600, 2048), (2048, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf105 = empty_strided_cuda((600, 2048), (2048, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_2, key_6, value_2, value_6], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf45, buf48, buf102, buf105, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_2], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf46 = torch.ops.aten._weight_int4pack_mm.default(buf45, arg38_1, 128, arg39_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg38_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg39_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf47 = buf46
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf46
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_2], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf49 = torch.ops.aten._weight_int4pack_mm.default(buf48, arg41_1, 128, arg42_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg41_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg42_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf50 = buf49
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf49
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf51 = reinterpret_tensor(buf39, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf39  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_11], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf51, arg37_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg37_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf52 = reinterpret_tensor(buf47, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf47  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_11], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf52, arg40_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg40_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf53 = reinterpret_tensor(buf50, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf50  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_11], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf53, arg43_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg43_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf54 = empty_strided_cuda((2, 16, 1, 304), (4864, 304, 304, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf111 = empty_strided_cuda((2, 16, 1, 304), (4864, 304, 304, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_11, hidden_states_30], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf54, buf111, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_11], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf55 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf51, buf52, buf53, reinterpret_tensor(buf54, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf51
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf52
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf53
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf56 = buf55[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf55
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf60 = buf37; del buf37  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_14], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf56, buf60, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_14], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf61 = torch.ops.aten._weight_int4pack_mm.default(buf60, arg44_1, 128, arg45_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg44_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg45_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf62 = buf61
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf61
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf66 = reinterpret_tensor(buf56, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf56  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_6, hidden_states_16, hidden_states_17, mul_6, norm_hidden_states_2, norm_hidden_states_3], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf62, arg46_1, buf36, arg22_1, buf9, arg14_1, buf66, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf67 = buf60; del buf60  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_18], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf66, buf67, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_18], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf68 = torch.ops.aten._weight_int4pack_mm.default(buf67, arg47_1, 128, arg48_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg47_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg48_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf69 = buf68
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf68
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf70 = empty_strided_cuda((8192, 5120), (5120, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_21], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf69, arg49_1, buf70, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg49_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf69
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_21], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf71 = torch.ops.aten._weight_int4pack_mm.default(buf70, arg50_1, 128, arg51_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg50_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg51_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf72 = buf71
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf71
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf73 = reinterpret_tensor(buf72, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf72  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf77 = buf66; del buf66  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_10, ff_output, hidden_states_16, hidden_states_17, hidden_states_22, mul_8, norm_hidden_states_4, norm_hidden_states_5], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf73, arg22_1, buf9, arg14_1, arg52_1, buf62, arg46_1, buf36, arg53_1, buf77, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg22_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg46_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg52_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf36
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf62
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf78 = buf67; del buf67  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf81 = buf19; del buf19  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf84 = buf16; del buf16  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_4, query_4, value_4], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf77, buf78, buf81, buf84, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf77
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_4], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf79 = torch.ops.aten._weight_int4pack_mm.default(buf78, arg54_1, 128, arg55_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg54_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg55_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf80 = buf79
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf79
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_4], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf82 = torch.ops.aten._weight_int4pack_mm.default(buf81, arg57_1, 128, arg58_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg57_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg58_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf83 = buf82
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf82
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_4], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf85 = torch.ops.aten._weight_int4pack_mm.default(buf84, arg60_1, 128, arg61_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg60_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg61_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf86 = buf85
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf85
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf87 = reinterpret_tensor(buf80, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf80  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_23], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf87, arg56_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg56_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf88 = reinterpret_tensor(buf83, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf83  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_23], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf88, arg59_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg59_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf89 = reinterpret_tensor(buf86, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf86  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_23], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf89, arg62_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg62_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_23], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf90 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf87, buf88, buf89, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf87
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf88
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf89
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf91 = buf90[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf90
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf95 = buf84; del buf84  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_26], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf91, buf95, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf91
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_26], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf96 = torch.ops.aten._weight_int4pack_mm.default(buf95, arg63_1, 128, arg64_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg63_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg64_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf97 = buf96
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf96
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf98 = reinterpret_tensor(buf97, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf97  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_1, hidden_states_28, hidden_states_29], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf98, arg53_1, buf9, arg14_1, arg65_1, buf73, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg65_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf73
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf99 = buf95; del buf95  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_6], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf98, buf99, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_6], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf100 = torch.ops.aten._weight_int4pack_mm.default(buf99, arg66_1, 128, arg67_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg66_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg67_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf101 = buf100
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf100
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_6], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf103 = torch.ops.aten._weight_int4pack_mm.default(buf102, arg69_1, 128, arg70_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg69_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg70_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf104 = buf103
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf103
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_6], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf106 = torch.ops.aten._weight_int4pack_mm.default(buf105, arg72_1, 128, arg73_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg72_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg73_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf107 = buf106
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf106
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf108 = reinterpret_tensor(buf101, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf101  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_30], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf108, arg68_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg68_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf109 = reinterpret_tensor(buf104, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf104  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_30], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf109, arg71_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg71_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf110 = reinterpret_tensor(buf107, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf107  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_30], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf110, arg74_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg74_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_30], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf112 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf108, buf109, buf110, reinterpret_tensor(buf111, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf108
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf109
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf110
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf113 = buf112[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf112
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf117 = buf99; del buf99  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_33], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf113, buf117, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_33], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf118 = torch.ops.aten._weight_int4pack_mm.default(buf117, arg75_1, 128, arg76_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg75_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg76_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf119 = buf118
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf118
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf123 = reinterpret_tensor(buf113, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf113  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_14, hidden_states_35, hidden_states_36, mul_10, norm_hidden_states_6, norm_hidden_states_7], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf119, arg77_1, buf98, arg53_1, buf9, arg14_1, buf123, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf124 = buf117; del buf117  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_37], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf123, buf124, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_37], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf125 = torch.ops.aten._weight_int4pack_mm.default(buf124, arg78_1, 128, arg79_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg78_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg79_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf126 = buf125
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf125
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf127 = buf70; del buf70  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_40], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf126, arg80_1, buf127, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg80_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf126
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_40], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf128 = torch.ops.aten._weight_int4pack_mm.default(buf127, arg81_1, 128, arg82_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg81_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg82_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf129 = buf128
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf128
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf130 = reinterpret_tensor(buf129, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf129  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf134 = buf123; del buf123  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_18, ff_output_1, hidden_states_35, hidden_states_36, hidden_states_41, mul_12, norm_hidden_states_8, norm_hidden_states_9], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf130, arg53_1, buf9, arg14_1, arg83_1, buf119, arg77_1, buf98, arg84_1, buf134, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg53_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg77_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg83_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf119
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf98
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf135 = buf124; del buf124  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf138 = buf81; del buf81  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf141 = buf78; del buf78  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_8, query_8, value_8], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf134, buf135, buf138, buf141, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf134
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_8], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf136 = torch.ops.aten._weight_int4pack_mm.default(buf135, arg85_1, 128, arg86_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg85_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg86_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf137 = buf136
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf136
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_8], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf139 = torch.ops.aten._weight_int4pack_mm.default(buf138, arg88_1, 128, arg89_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg88_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg89_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf140 = buf139
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf139
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_8], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf142 = torch.ops.aten._weight_int4pack_mm.default(buf141, arg91_1, 128, arg92_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg91_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg92_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf143 = buf142
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf142
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf144 = reinterpret_tensor(buf137, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf137  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_42], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf144, arg87_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg87_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf145 = reinterpret_tensor(buf140, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf140  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_42], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf145, arg90_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg90_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf146 = reinterpret_tensor(buf143, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf143  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_42], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf146, arg93_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg93_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_42], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf147 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf144, buf145, buf146, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf144
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf145
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf146
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf148 = buf147[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf147
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf152 = buf141; del buf141  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_45], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf148, buf152, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf148
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_45], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf153 = torch.ops.aten._weight_int4pack_mm.default(buf152, arg94_1, 128, arg95_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg94_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg95_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf154 = buf153
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf153
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf155 = reinterpret_tensor(buf154, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf154  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_2, hidden_states_47, hidden_states_48], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf155, arg84_1, buf9, arg14_1, arg96_1, buf130, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg96_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf130
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf156 = buf152; del buf152  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_10], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf155, buf156, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_10], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf157 = torch.ops.aten._weight_int4pack_mm.default(buf156, arg97_1, 128, arg98_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg97_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg98_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf158 = buf157
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf157
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf159 = buf105; del buf105  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf162 = buf102; del buf102  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf216 = buf48; del buf48  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf219 = buf45; del buf45  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_10, key_14, value_10, value_14], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf159, buf162, buf216, buf219, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_10], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf160 = torch.ops.aten._weight_int4pack_mm.default(buf159, arg100_1, 128, arg101_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg100_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg101_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf161 = buf160
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf160
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_10], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf163 = torch.ops.aten._weight_int4pack_mm.default(buf162, arg103_1, 128, arg104_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg103_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg104_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf164 = buf163
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf163
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf165 = reinterpret_tensor(buf158, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf158  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_49], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf165, arg99_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg99_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf166 = reinterpret_tensor(buf161, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf161  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_49], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf166, arg102_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg102_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf167 = reinterpret_tensor(buf164, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf164  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_49], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf167, arg105_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg105_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf168 = buf111; del buf111  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf225 = buf54; del buf54  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_49, hidden_states_68], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf168, buf225, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_49], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf169 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf165, buf166, buf167, reinterpret_tensor(buf168, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf165
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf166
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf167
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf170 = buf169[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf169
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf174 = buf156; del buf156  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_52], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf170, buf174, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_52], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf175 = torch.ops.aten._weight_int4pack_mm.default(buf174, arg106_1, 128, arg107_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg106_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg107_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf176 = buf175
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf175
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf180 = reinterpret_tensor(buf170, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf170  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_22, hidden_states_54, hidden_states_55, mul_14, norm_hidden_states_10, norm_hidden_states_11], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf176, arg108_1, buf155, arg84_1, buf9, arg14_1, buf180, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf181 = buf174; del buf174  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_56], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf180, buf181, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_56], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf182 = torch.ops.aten._weight_int4pack_mm.default(buf181, arg109_1, 128, arg110_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg109_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg110_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf183 = buf182
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf182
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf184 = buf127; del buf127  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_59], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf183, arg111_1, buf184, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg111_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf183
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_59], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf185 = torch.ops.aten._weight_int4pack_mm.default(buf184, arg112_1, 128, arg113_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg112_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg113_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf186 = buf185
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf185
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf187 = reinterpret_tensor(buf186, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf186  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf191 = buf180; del buf180  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_26, ff_output_2, hidden_states_54, hidden_states_55, hidden_states_60, mul_16, norm_hidden_states_12, norm_hidden_states_13], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf187, arg84_1, buf9, arg14_1, arg114_1, buf176, arg108_1, buf155, arg115_1, buf191, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg108_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg114_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg84_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf155
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf176
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf192 = buf181; del buf181  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf195 = buf138; del buf138  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf198 = buf135; del buf135  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_12, query_12, value_12], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf191, buf192, buf195, buf198, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf191
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_12], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf193 = torch.ops.aten._weight_int4pack_mm.default(buf192, arg116_1, 128, arg117_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg116_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg117_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf194 = buf193
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf193
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_12], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf196 = torch.ops.aten._weight_int4pack_mm.default(buf195, arg119_1, 128, arg120_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg119_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg120_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf197 = buf196
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf196
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_12], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf199 = torch.ops.aten._weight_int4pack_mm.default(buf198, arg122_1, 128, arg123_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg122_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg123_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf200 = buf199
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf199
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf201 = reinterpret_tensor(buf194, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf194  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_61], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf201, arg118_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg118_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf202 = reinterpret_tensor(buf197, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf197  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_61], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf202, arg121_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg121_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf203 = reinterpret_tensor(buf200, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf200  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_61], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf203, arg124_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg124_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_61], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf204 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf201, buf202, buf203, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf201
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf202
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf203
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf205 = buf204[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf204
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf209 = buf198; del buf198  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_64], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf205, buf209, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf205
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_64], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf210 = torch.ops.aten._weight_int4pack_mm.default(buf209, arg125_1, 128, arg126_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg125_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg126_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf211 = buf210
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf210
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf212 = reinterpret_tensor(buf211, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf211  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_3, hidden_states_66, hidden_states_67], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf212, arg115_1, buf9, arg14_1, arg127_1, buf187, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg127_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf187
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf213 = buf209; del buf209  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_14], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf212, buf213, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_14], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf214 = torch.ops.aten._weight_int4pack_mm.default(buf213, arg128_1, 128, arg129_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg128_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg129_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf215 = buf214
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf214
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_14], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf217 = torch.ops.aten._weight_int4pack_mm.default(buf216, arg131_1, 128, arg132_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg131_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg132_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf218 = buf217
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf217
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_14], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf220 = torch.ops.aten._weight_int4pack_mm.default(buf219, arg134_1, 128, arg135_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg134_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg135_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf221 = buf220
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf220
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf222 = reinterpret_tensor(buf215, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf215  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_68], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf222, arg130_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg130_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf223 = reinterpret_tensor(buf218, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf218  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_68], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf223, arg133_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg133_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf224 = reinterpret_tensor(buf221, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf221  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_68], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf224, arg136_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg136_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_68], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf226 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf222, buf223, buf224, reinterpret_tensor(buf225, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf222
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf223
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf224
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf227 = buf226[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf226
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf231 = buf213; del buf213  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_71], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf227, buf231, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_71], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf232 = torch.ops.aten._weight_int4pack_mm.default(buf231, arg137_1, 128, arg138_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg137_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg138_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf233 = buf232
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf232
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf237 = reinterpret_tensor(buf227, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf227  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_30, hidden_states_73, hidden_states_74, mul_18, norm_hidden_states_14, norm_hidden_states_15], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf233, arg139_1, buf212, arg115_1, buf9, arg14_1, buf237, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf238 = buf231; del buf231  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_75], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf237, buf238, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_75], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf239 = torch.ops.aten._weight_int4pack_mm.default(buf238, arg140_1, 128, arg141_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg140_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg141_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf240 = buf239
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf239
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf241 = buf184; del buf184  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_78], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf240, arg142_1, buf241, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg142_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf240
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_78], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf242 = torch.ops.aten._weight_int4pack_mm.default(buf241, arg143_1, 128, arg144_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg143_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg144_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf243 = buf242
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf242
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf244 = reinterpret_tensor(buf243, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf243  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf248 = buf237; del buf237  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_34, ff_output_3, hidden_states_73, hidden_states_74, hidden_states_79, mul_20, norm_hidden_states_16, norm_hidden_states_17], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf244, arg115_1, buf9, arg14_1, arg145_1, buf233, arg139_1, buf212, arg146_1, buf248, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg115_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg139_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg145_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf212
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf233
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf249 = buf238; del buf238  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf252 = buf195; del buf195  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf255 = buf192; del buf192  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_16, query_16, value_16], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf248, buf249, buf252, buf255, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf248
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_16], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf250 = torch.ops.aten._weight_int4pack_mm.default(buf249, arg147_1, 128, arg148_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg147_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg148_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf251 = buf250
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf250
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_16], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf253 = torch.ops.aten._weight_int4pack_mm.default(buf252, arg150_1, 128, arg151_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg150_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg151_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf254 = buf253
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf253
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_16], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf256 = torch.ops.aten._weight_int4pack_mm.default(buf255, arg153_1, 128, arg154_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg153_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg154_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf257 = buf256
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf256
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf258 = reinterpret_tensor(buf251, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf251  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_80], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf258, arg149_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg149_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf259 = reinterpret_tensor(buf254, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf254  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_80], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf259, arg152_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg152_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf260 = reinterpret_tensor(buf257, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf257  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_80], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf260, arg155_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg155_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_80], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf261 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf258, buf259, buf260, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf258
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf259
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf260
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf262 = buf261[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf261
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf266 = buf255; del buf255  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_83], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf262, buf266, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf262
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_83], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf267 = torch.ops.aten._weight_int4pack_mm.default(buf266, arg156_1, 128, arg157_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg156_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg157_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf268 = buf267
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf267
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf269 = reinterpret_tensor(buf268, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf268  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_4, hidden_states_85, hidden_states_86], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf269, arg146_1, buf9, arg14_1, arg158_1, buf244, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg158_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf244
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf270 = buf266; del buf266  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_18], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf269, buf270, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_18], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf271 = torch.ops.aten._weight_int4pack_mm.default(buf270, arg159_1, 128, arg160_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg159_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg160_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf272 = buf271
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf271
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf273 = buf219; del buf219  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf276 = buf216; del buf216  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf330 = buf162; del buf162  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf333 = buf159; del buf159  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_18, key_22, value_18, value_22], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf273, buf276, buf330, buf333, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_18], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf274 = torch.ops.aten._weight_int4pack_mm.default(buf273, arg162_1, 128, arg163_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg162_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg163_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf275 = buf274
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf274
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_18], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf277 = torch.ops.aten._weight_int4pack_mm.default(buf276, arg165_1, 128, arg166_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg165_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg166_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf278 = buf277
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf277
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf279 = reinterpret_tensor(buf272, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf272  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_87], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf279, arg161_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg161_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf280 = reinterpret_tensor(buf275, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf275  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_87], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf280, arg164_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg164_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf281 = reinterpret_tensor(buf278, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf278  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_87], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf281, arg167_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg167_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf282 = buf225; del buf225  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf339 = buf168; del buf168  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_106, hidden_states_87], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf282, buf339, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_87], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf283 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf279, buf280, buf281, reinterpret_tensor(buf282, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf279
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf280
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf281
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf284 = buf283[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf283
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf288 = buf270; del buf270  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_90], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf284, buf288, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_90], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf289 = torch.ops.aten._weight_int4pack_mm.default(buf288, arg168_1, 128, arg169_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg168_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg169_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf290 = buf289
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf289
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf294 = reinterpret_tensor(buf284, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf284  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_38, hidden_states_92, hidden_states_93, mul_22, norm_hidden_states_18, norm_hidden_states_19], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf290, arg170_1, buf269, arg146_1, buf9, arg14_1, buf294, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf295 = buf288; del buf288  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_94], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf294, buf295, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_94], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf296 = torch.ops.aten._weight_int4pack_mm.default(buf295, arg171_1, 128, arg172_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg171_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg172_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf297 = buf296
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf296
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf298 = buf241; del buf241  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_97], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf297, arg173_1, buf298, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg173_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf297
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_97], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf299 = torch.ops.aten._weight_int4pack_mm.default(buf298, arg174_1, 128, arg175_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg174_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg175_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf300 = buf299
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf299
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf301 = reinterpret_tensor(buf300, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf300  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf305 = buf294; del buf294  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_42, ff_output_4, hidden_states_92, hidden_states_93, hidden_states_98, mul_24, norm_hidden_states_20, norm_hidden_states_21], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf301, arg146_1, buf9, arg14_1, arg176_1, buf290, arg170_1, buf269, arg177_1, buf305, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg146_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg170_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg176_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf269
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf290
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf306 = buf295; del buf295  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf309 = buf252; del buf252  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf312 = buf249; del buf249  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_20, query_20, value_20], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf305, buf306, buf309, buf312, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf305
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_20], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf307 = torch.ops.aten._weight_int4pack_mm.default(buf306, arg178_1, 128, arg179_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg178_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg179_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf308 = buf307
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf307
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_20], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf310 = torch.ops.aten._weight_int4pack_mm.default(buf309, arg181_1, 128, arg182_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg181_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg182_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf311 = buf310
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf310
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_20], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf313 = torch.ops.aten._weight_int4pack_mm.default(buf312, arg184_1, 128, arg185_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg184_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg185_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf314 = buf313
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf313
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf315 = reinterpret_tensor(buf308, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf308  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_99], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf315, arg180_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg180_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf316 = reinterpret_tensor(buf311, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf311  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_99], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf316, arg183_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg183_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf317 = reinterpret_tensor(buf314, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf314  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_99], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf317, arg186_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg186_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_99], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf318 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf315, buf316, buf317, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf315
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf316
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf317
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf319 = buf318[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf318
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf323 = buf312; del buf312  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_102], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf319, buf323, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf319
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_102], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf324 = torch.ops.aten._weight_int4pack_mm.default(buf323, arg187_1, 128, arg188_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg187_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg188_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf325 = buf324
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf324
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf326 = reinterpret_tensor(buf325, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf325  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_5, hidden_states_104, hidden_states_105], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf326, arg177_1, buf9, arg14_1, arg189_1, buf301, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg189_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf301
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf327 = buf323; del buf323  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_22], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf326, buf327, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_22], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf328 = torch.ops.aten._weight_int4pack_mm.default(buf327, arg190_1, 128, arg191_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg190_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg191_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf329 = buf328
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf328
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_22], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf331 = torch.ops.aten._weight_int4pack_mm.default(buf330, arg193_1, 128, arg194_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg193_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg194_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf332 = buf331
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf331
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_22], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf334 = torch.ops.aten._weight_int4pack_mm.default(buf333, arg196_1, 128, arg197_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg196_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg197_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf335 = buf334
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf334
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf336 = reinterpret_tensor(buf329, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf329  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_106], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf336, arg192_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg192_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf337 = reinterpret_tensor(buf332, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf332  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_106], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf337, arg195_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg195_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf338 = reinterpret_tensor(buf335, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf335  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_106], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf338, arg198_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg198_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_106], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf340 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf336, buf337, buf338, reinterpret_tensor(buf339, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf336
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf337
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf338
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf341 = buf340[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf340
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf345 = buf327; del buf327  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_109], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf341, buf345, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_109], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf346 = torch.ops.aten._weight_int4pack_mm.default(buf345, arg199_1, 128, arg200_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg199_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg200_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf347 = buf346
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf346
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf351 = reinterpret_tensor(buf341, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf341  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_46, hidden_states_111, hidden_states_112, mul_26, norm_hidden_states_22, norm_hidden_states_23], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf347, arg201_1, buf326, arg177_1, buf9, arg14_1, buf351, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf352 = buf345; del buf345  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_113], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf351, buf352, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_113], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf353 = torch.ops.aten._weight_int4pack_mm.default(buf352, arg202_1, 128, arg203_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg202_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg203_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf354 = buf353
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf353
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf355 = buf298; del buf298  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_116], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf354, arg204_1, buf355, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg204_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf354
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_116], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf356 = torch.ops.aten._weight_int4pack_mm.default(buf355, arg205_1, 128, arg206_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg205_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg206_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf357 = buf356
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf356
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf358 = reinterpret_tensor(buf357, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf357  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf362 = buf351; del buf351  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_50, ff_output_5, hidden_states_111, hidden_states_112, hidden_states_117, mul_28, norm_hidden_states_24, norm_hidden_states_25], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf358, arg177_1, buf9, arg14_1, arg207_1, buf347, arg201_1, buf326, arg208_1, buf362, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg177_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg201_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg207_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf326
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf347
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf363 = buf352; del buf352  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf366 = buf309; del buf309  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf369 = buf306; del buf306  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_24, query_24, value_24], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf362, buf363, buf366, buf369, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf362
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_24], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf364 = torch.ops.aten._weight_int4pack_mm.default(buf363, arg209_1, 128, arg210_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg209_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg210_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf365 = buf364
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf364
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_24], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf367 = torch.ops.aten._weight_int4pack_mm.default(buf366, arg212_1, 128, arg213_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg212_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg213_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf368 = buf367
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf367
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_24], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf370 = torch.ops.aten._weight_int4pack_mm.default(buf369, arg215_1, 128, arg216_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg215_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg216_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf371 = buf370
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf370
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf372 = reinterpret_tensor(buf365, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf365  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_118], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf372, arg211_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg211_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf373 = reinterpret_tensor(buf368, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf368  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_118], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf373, arg214_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg214_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf374 = reinterpret_tensor(buf371, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf371  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_118], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf374, arg217_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg217_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_118], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf375 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf372, buf373, buf374, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf372
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf373
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf374
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf376 = buf375[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf375
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf380 = buf369; del buf369  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_121], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf376, buf380, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf376
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_121], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf381 = torch.ops.aten._weight_int4pack_mm.default(buf380, arg218_1, 128, arg219_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg218_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg219_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf382 = buf381
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf381
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf383 = reinterpret_tensor(buf382, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf382  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_6, hidden_states_123, hidden_states_124], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf383, arg208_1, buf9, arg14_1, arg220_1, buf358, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg220_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf358
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf384 = buf380; del buf380  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_26], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf383, buf384, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_26], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf385 = torch.ops.aten._weight_int4pack_mm.default(buf384, arg221_1, 128, arg222_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg221_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg222_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf386 = buf385
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf385
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf387 = buf333; del buf333  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf390 = buf330; del buf330  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf444 = buf276; del buf276  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf447 = buf273; del buf273  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_26, key_30, value_26, value_30], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf387, buf390, buf444, buf447, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_26], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf388 = torch.ops.aten._weight_int4pack_mm.default(buf387, arg224_1, 128, arg225_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg224_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg225_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf389 = buf388
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf388
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_26], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf391 = torch.ops.aten._weight_int4pack_mm.default(buf390, arg227_1, 128, arg228_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg227_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg228_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf392 = buf391
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf391
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf393 = reinterpret_tensor(buf386, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf386  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_125], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf393, arg223_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg223_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf394 = reinterpret_tensor(buf389, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf389  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_125], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf394, arg226_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg226_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf395 = reinterpret_tensor(buf392, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf392  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_125], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf395, arg229_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg229_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf396 = buf339; del buf339  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf453 = buf282; del buf282  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_125, hidden_states_144], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf396, buf453, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_125], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf397 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf393, buf394, buf395, reinterpret_tensor(buf396, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf393
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf394
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf395
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf398 = buf397[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf397
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf402 = buf384; del buf384  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_128], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf398, buf402, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_128], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf403 = torch.ops.aten._weight_int4pack_mm.default(buf402, arg230_1, 128, arg231_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg230_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg231_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf404 = buf403
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf403
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf408 = reinterpret_tensor(buf398, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf398  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_54, hidden_states_130, hidden_states_131, mul_30, norm_hidden_states_26, norm_hidden_states_27], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf404, arg232_1, buf383, arg208_1, buf9, arg14_1, buf408, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf409 = buf402; del buf402  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_132], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf408, buf409, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_132], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf410 = torch.ops.aten._weight_int4pack_mm.default(buf409, arg233_1, 128, arg234_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg233_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg234_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf411 = buf410
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf410
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf412 = buf355; del buf355  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_135], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf411, arg235_1, buf412, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg235_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf411
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_135], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf413 = torch.ops.aten._weight_int4pack_mm.default(buf412, arg236_1, 128, arg237_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg236_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg237_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf414 = buf413
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf413
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf415 = reinterpret_tensor(buf414, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf414  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf419 = buf408; del buf408  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_58, ff_output_6, hidden_states_130, hidden_states_131, hidden_states_136, mul_32, norm_hidden_states_28, norm_hidden_states_29], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf415, arg208_1, buf9, arg14_1, arg238_1, buf404, arg232_1, buf383, arg239_1, buf419, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg208_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg232_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg238_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf383
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf404
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf420 = buf409; del buf409  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf423 = buf366; del buf366  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf426 = buf363; del buf363  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_28, query_28, value_28], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf419, buf420, buf423, buf426, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf419
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_28], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf421 = torch.ops.aten._weight_int4pack_mm.default(buf420, arg240_1, 128, arg241_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg240_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg241_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf422 = buf421
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf421
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_28], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf424 = torch.ops.aten._weight_int4pack_mm.default(buf423, arg243_1, 128, arg244_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg243_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg244_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf425 = buf424
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf424
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_28], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf427 = torch.ops.aten._weight_int4pack_mm.default(buf426, arg246_1, 128, arg247_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg246_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg247_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf428 = buf427
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf427
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf429 = reinterpret_tensor(buf422, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf422  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_137], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf429, arg242_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg242_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf430 = reinterpret_tensor(buf425, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf425  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_137], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf430, arg245_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg245_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf431 = reinterpret_tensor(buf428, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf428  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_137], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf431, arg248_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg248_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_137], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf432 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf429, buf430, buf431, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf429
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf430
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf431
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf433 = buf432[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf432
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf437 = buf426; del buf426  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_140], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf433, buf437, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf433
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_140], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf438 = torch.ops.aten._weight_int4pack_mm.default(buf437, arg249_1, 128, arg250_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg249_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg250_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf439 = buf438
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf438
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf440 = reinterpret_tensor(buf439, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf439  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_7, hidden_states_142, hidden_states_143], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf440, arg239_1, buf9, arg14_1, arg251_1, buf415, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg251_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf415
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf441 = buf437; del buf437  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_30], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf440, buf441, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_30], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf442 = torch.ops.aten._weight_int4pack_mm.default(buf441, arg252_1, 128, arg253_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg252_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg253_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf443 = buf442
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf442
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_30], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf445 = torch.ops.aten._weight_int4pack_mm.default(buf444, arg255_1, 128, arg256_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg255_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg256_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf446 = buf445
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf445
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_30], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf448 = torch.ops.aten._weight_int4pack_mm.default(buf447, arg258_1, 128, arg259_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg258_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg259_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf449 = buf448
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf448
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf450 = reinterpret_tensor(buf443, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf443  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_144], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf450, arg254_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg254_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf451 = reinterpret_tensor(buf446, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf446  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_144], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf451, arg257_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg257_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf452 = reinterpret_tensor(buf449, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf449  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_144], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf452, arg260_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg260_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_144], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf454 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf450, buf451, buf452, reinterpret_tensor(buf453, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf450
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf451
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf452
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf455 = buf454[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf454
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf459 = buf441; del buf441  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_147], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf455, buf459, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_147], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf460 = torch.ops.aten._weight_int4pack_mm.default(buf459, arg261_1, 128, arg262_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg261_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg262_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf461 = buf460
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf460
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf465 = reinterpret_tensor(buf455, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf455  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_62, hidden_states_149, hidden_states_150, mul_34, norm_hidden_states_30, norm_hidden_states_31], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf461, arg263_1, buf440, arg239_1, buf9, arg14_1, buf465, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf466 = buf459; del buf459  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_151], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf465, buf466, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_151], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf467 = torch.ops.aten._weight_int4pack_mm.default(buf466, arg264_1, 128, arg265_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg264_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg265_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf468 = buf467
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf467
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf469 = buf412; del buf412  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_154], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf468, arg266_1, buf469, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg266_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf468
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_154], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf470 = torch.ops.aten._weight_int4pack_mm.default(buf469, arg267_1, 128, arg268_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg267_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg268_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf471 = buf470
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf470
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf472 = reinterpret_tensor(buf471, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf471  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf476 = buf465; del buf465  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_66, ff_output_7, hidden_states_149, hidden_states_150, hidden_states_155, mul_36, norm_hidden_states_32, norm_hidden_states_33], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf472, arg239_1, buf9, arg14_1, arg269_1, buf461, arg263_1, buf440, arg270_1, buf476, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg239_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg263_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg269_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf440
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf461
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf477 = buf466; del buf466  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf480 = buf423; del buf423  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf483 = buf420; del buf420  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_32, query_32, value_32], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf476, buf477, buf480, buf483, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf476
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_32], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf478 = torch.ops.aten._weight_int4pack_mm.default(buf477, arg271_1, 128, arg272_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg271_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg272_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf479 = buf478
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf478
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_32], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf481 = torch.ops.aten._weight_int4pack_mm.default(buf480, arg274_1, 128, arg275_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg274_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg275_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf482 = buf481
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf481
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_32], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf484 = torch.ops.aten._weight_int4pack_mm.default(buf483, arg277_1, 128, arg278_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg277_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg278_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf485 = buf484
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf484
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf486 = reinterpret_tensor(buf479, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf479  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_156], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf486, arg273_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg273_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf487 = reinterpret_tensor(buf482, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf482  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_156], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf487, arg276_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg276_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf488 = reinterpret_tensor(buf485, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf485  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_156], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf488, arg279_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg279_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_156], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf489 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf486, buf487, buf488, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf486
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf487
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf488
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf490 = buf489[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf489
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf494 = buf483; del buf483  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_159], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf490, buf494, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf490
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_159], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf495 = torch.ops.aten._weight_int4pack_mm.default(buf494, arg280_1, 128, arg281_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg280_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg281_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf496 = buf495
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf495
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf497 = reinterpret_tensor(buf496, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf496  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_8, hidden_states_161, hidden_states_162], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf497, arg270_1, buf9, arg14_1, arg282_1, buf472, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg282_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf472
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf498 = buf494; del buf494  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_34], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf497, buf498, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_34], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf499 = torch.ops.aten._weight_int4pack_mm.default(buf498, arg283_1, 128, arg284_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg283_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg284_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf500 = buf499
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf499
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf501 = buf447; del buf447  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf504 = buf444; del buf444  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf558 = buf390; del buf390  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf561 = buf387; del buf387  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_34, key_38, value_34, value_38], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf501, buf504, buf558, buf561, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_34], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf502 = torch.ops.aten._weight_int4pack_mm.default(buf501, arg286_1, 128, arg287_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg286_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg287_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf503 = buf502
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf502
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_34], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf505 = torch.ops.aten._weight_int4pack_mm.default(buf504, arg289_1, 128, arg290_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg289_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg290_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf506 = buf505
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf505
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf507 = reinterpret_tensor(buf500, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf500  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_163], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf507, arg285_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg285_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf508 = reinterpret_tensor(buf503, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf503  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_163], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf508, arg288_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg288_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf509 = reinterpret_tensor(buf506, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf506  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_163], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf509, arg291_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg291_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf510 = buf453; del buf453  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf567 = buf396; del buf396  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_163, hidden_states_182], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf510, buf567, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_163], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf511 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf507, buf508, buf509, reinterpret_tensor(buf510, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf507
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf508
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf509
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf512 = buf511[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf511
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf516 = buf498; del buf498  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_166], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf512, buf516, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_166], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf517 = torch.ops.aten._weight_int4pack_mm.default(buf516, arg292_1, 128, arg293_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg292_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg293_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf518 = buf517
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf517
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf522 = reinterpret_tensor(buf512, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf512  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_70, hidden_states_168, hidden_states_169, mul_38, norm_hidden_states_34, norm_hidden_states_35], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf518, arg294_1, buf497, arg270_1, buf9, arg14_1, buf522, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf523 = buf516; del buf516  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_170], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf522, buf523, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_170], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf524 = torch.ops.aten._weight_int4pack_mm.default(buf523, arg295_1, 128, arg296_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg295_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg296_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf525 = buf524
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf524
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf526 = buf469; del buf469  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_173], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf525, arg297_1, buf526, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg297_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf525
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_173], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf527 = torch.ops.aten._weight_int4pack_mm.default(buf526, arg298_1, 128, arg299_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg298_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg299_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf528 = buf527
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf527
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf529 = reinterpret_tensor(buf528, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf528  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf533 = buf522; del buf522  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_74, ff_output_8, hidden_states_168, hidden_states_169, hidden_states_174, mul_40, norm_hidden_states_36, norm_hidden_states_37], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf529, arg270_1, buf9, arg14_1, arg300_1, buf518, arg294_1, buf497, arg301_1, buf533, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg270_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg294_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg300_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf497
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf518
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf534 = buf523; del buf523  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf537 = buf480; del buf480  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf540 = buf477; del buf477  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_36, query_36, value_36], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf533, buf534, buf537, buf540, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf533
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_36], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf535 = torch.ops.aten._weight_int4pack_mm.default(buf534, arg302_1, 128, arg303_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg302_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg303_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf536 = buf535
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf535
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_36], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf538 = torch.ops.aten._weight_int4pack_mm.default(buf537, arg305_1, 128, arg306_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg305_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg306_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf539 = buf538
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf538
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_36], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf541 = torch.ops.aten._weight_int4pack_mm.default(buf540, arg308_1, 128, arg309_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg308_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg309_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf542 = buf541
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf541
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf543 = reinterpret_tensor(buf536, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf536  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_175], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf543, arg304_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg304_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf544 = reinterpret_tensor(buf539, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf539  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_175], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf544, arg307_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg307_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf545 = reinterpret_tensor(buf542, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf542  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_175], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf545, arg310_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg310_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_175], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf546 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf543, buf544, buf545, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf543
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf544
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf545
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf547 = buf546[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf546
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf551 = buf540; del buf540  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_178], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf547, buf551, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf547
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_178], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf552 = torch.ops.aten._weight_int4pack_mm.default(buf551, arg311_1, 128, arg312_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg311_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg312_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf553 = buf552
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf552
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf554 = reinterpret_tensor(buf553, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf553  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_9, hidden_states_180, hidden_states_181], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf554, arg301_1, buf9, arg14_1, arg313_1, buf529, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg313_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf529
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf555 = buf551; del buf551  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_38], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf554, buf555, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_38], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf556 = torch.ops.aten._weight_int4pack_mm.default(buf555, arg314_1, 128, arg315_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg314_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg315_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf557 = buf556
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf556
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_38], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf559 = torch.ops.aten._weight_int4pack_mm.default(buf558, arg317_1, 128, arg318_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg317_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg318_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf560 = buf559
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf559
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_38], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf562 = torch.ops.aten._weight_int4pack_mm.default(buf561, arg320_1, 128, arg321_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg320_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg321_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf563 = buf562
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf562
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf564 = reinterpret_tensor(buf557, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf557  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_182], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf564, arg316_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg316_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf565 = reinterpret_tensor(buf560, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf560  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_182], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf565, arg319_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg319_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf566 = reinterpret_tensor(buf563, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf563  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_182], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf566, arg322_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg322_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_182], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf568 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf564, buf565, buf566, reinterpret_tensor(buf567, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf564
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf565
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf566
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf569 = buf568[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf568
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf573 = buf555; del buf555  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_185], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf569, buf573, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_185], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf574 = torch.ops.aten._weight_int4pack_mm.default(buf573, arg323_1, 128, arg324_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg323_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg324_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf575 = buf574
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf574
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf579 = reinterpret_tensor(buf569, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf569  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_78, hidden_states_187, hidden_states_188, mul_42, norm_hidden_states_38, norm_hidden_states_39], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf575, arg325_1, buf554, arg301_1, buf9, arg14_1, buf579, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf580 = buf573; del buf573  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_189], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf579, buf580, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_189], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf581 = torch.ops.aten._weight_int4pack_mm.default(buf580, arg326_1, 128, arg327_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg326_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg327_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf582 = buf581
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf581
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf583 = buf526; del buf526  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_192], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf582, arg328_1, buf583, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg328_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf582
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_192], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf584 = torch.ops.aten._weight_int4pack_mm.default(buf583, arg329_1, 128, arg330_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg329_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg330_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf585 = buf584
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf584
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf586 = reinterpret_tensor(buf585, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf585  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf590 = buf579; del buf579  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_82, ff_output_9, hidden_states_187, hidden_states_188, hidden_states_193, mul_44, norm_hidden_states_40, norm_hidden_states_41], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf586, arg301_1, buf9, arg14_1, arg331_1, buf575, arg325_1, buf554, arg332_1, buf590, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg301_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg325_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg331_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf554
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf575
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf591 = buf580; del buf580  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf594 = buf537; del buf537  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf597 = buf534; del buf534  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_40, query_40, value_40], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf590, buf591, buf594, buf597, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf590
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_40], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf592 = torch.ops.aten._weight_int4pack_mm.default(buf591, arg333_1, 128, arg334_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg333_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg334_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf593 = buf592
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf592
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_40], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf595 = torch.ops.aten._weight_int4pack_mm.default(buf594, arg336_1, 128, arg337_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg336_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg337_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf596 = buf595
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf595
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_40], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf598 = torch.ops.aten._weight_int4pack_mm.default(buf597, arg339_1, 128, arg340_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg339_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg340_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf599 = buf598
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf598
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf600 = reinterpret_tensor(buf593, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf593  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_194], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf600, arg335_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg335_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf601 = reinterpret_tensor(buf596, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf596  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_194], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf601, arg338_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg338_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf602 = reinterpret_tensor(buf599, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf599  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_194], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf602, arg341_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg341_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_194], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf603 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf600, buf601, buf602, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf600
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf601
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf602
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf604 = buf603[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf603
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf608 = buf597; del buf597  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_197], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf604, buf608, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf604
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_197], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf609 = torch.ops.aten._weight_int4pack_mm.default(buf608, arg342_1, 128, arg343_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg342_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg343_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf610 = buf609
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf609
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf611 = reinterpret_tensor(buf610, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf610  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_10, hidden_states_199, hidden_states_200], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf611, arg332_1, buf9, arg14_1, arg344_1, buf586, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg344_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf586
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf612 = buf608; del buf608  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_42], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf611, buf612, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_42], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf613 = torch.ops.aten._weight_int4pack_mm.default(buf612, arg345_1, 128, arg346_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg345_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg346_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf614 = buf613
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf613
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf615 = buf561; del buf561  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf618 = buf558; del buf558  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf672 = buf504; del buf504  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf675 = buf501; del buf501  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_42, key_46, value_42, value_46], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf615, buf618, buf672, buf675, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_42], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf616 = torch.ops.aten._weight_int4pack_mm.default(buf615, arg348_1, 128, arg349_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg348_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg349_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf617 = buf616
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf616
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_42], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf619 = torch.ops.aten._weight_int4pack_mm.default(buf618, arg351_1, 128, arg352_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg351_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg352_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf620 = buf619
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf619
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf621 = reinterpret_tensor(buf614, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf614  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_201], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf621, arg347_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg347_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf622 = reinterpret_tensor(buf617, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf617  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_201], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf622, arg350_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg350_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf623 = reinterpret_tensor(buf620, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf620  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_201], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf623, arg353_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg353_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf624 = buf567; del buf567  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf681 = buf510; del buf510  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_201, hidden_states_220], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf624, buf681, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_201], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf625 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf621, buf622, buf623, reinterpret_tensor(buf624, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf621
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf622
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf623
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf626 = buf625[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf625
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf630 = buf612; del buf612  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_204], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf626, buf630, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_204], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf631 = torch.ops.aten._weight_int4pack_mm.default(buf630, arg354_1, 128, arg355_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg354_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg355_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf632 = buf631
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf631
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf636 = reinterpret_tensor(buf626, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf626  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_86, hidden_states_206, hidden_states_207, mul_46, norm_hidden_states_42, norm_hidden_states_43], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf632, arg356_1, buf611, arg332_1, buf9, arg14_1, buf636, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf637 = buf630; del buf630  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_208], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf636, buf637, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_208], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf638 = torch.ops.aten._weight_int4pack_mm.default(buf637, arg357_1, 128, arg358_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg357_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg358_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf639 = buf638
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf638
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf640 = buf583; del buf583  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_211], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf639, arg359_1, buf640, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg359_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf639
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_211], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf641 = torch.ops.aten._weight_int4pack_mm.default(buf640, arg360_1, 128, arg361_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg360_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg361_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf642 = buf641
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf641
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf643 = reinterpret_tensor(buf642, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf642  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf647 = buf636; del buf636  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_90, ff_output_10, hidden_states_206, hidden_states_207, hidden_states_212, mul_48, norm_hidden_states_44, norm_hidden_states_45], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf643, arg332_1, buf9, arg14_1, arg362_1, buf632, arg356_1, buf611, arg363_1, buf647, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg332_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg356_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg362_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf611
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf632
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf648 = buf637; del buf637  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf651 = buf594; del buf594  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf654 = buf591; del buf591  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_44, query_44, value_44], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf647, buf648, buf651, buf654, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf647
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_44], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf649 = torch.ops.aten._weight_int4pack_mm.default(buf648, arg364_1, 128, arg365_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg364_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg365_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf650 = buf649
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf649
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_44], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf652 = torch.ops.aten._weight_int4pack_mm.default(buf651, arg367_1, 128, arg368_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg367_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg368_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf653 = buf652
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf652
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_44], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf655 = torch.ops.aten._weight_int4pack_mm.default(buf654, arg370_1, 128, arg371_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg370_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg371_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf656 = buf655
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf655
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf657 = reinterpret_tensor(buf650, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf650  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_213], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf657, arg366_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg366_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf658 = reinterpret_tensor(buf653, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf653  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_213], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf658, arg369_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg369_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf659 = reinterpret_tensor(buf656, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf656  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_213], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf659, arg372_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg372_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_213], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf660 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf657, buf658, buf659, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf657
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf658
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf659
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf661 = buf660[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf660
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf665 = buf654; del buf654  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_216], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf661, buf665, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf661
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_216], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf666 = torch.ops.aten._weight_int4pack_mm.default(buf665, arg373_1, 128, arg374_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg373_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg374_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf667 = buf666
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf666
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf668 = reinterpret_tensor(buf667, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf667  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_11, hidden_states_218, hidden_states_219], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf668, arg363_1, buf9, arg14_1, arg375_1, buf643, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg375_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf643
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf669 = buf665; del buf665  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_46], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf668, buf669, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_46], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf670 = torch.ops.aten._weight_int4pack_mm.default(buf669, arg376_1, 128, arg377_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg376_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg377_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf671 = buf670
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf670
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_46], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf673 = torch.ops.aten._weight_int4pack_mm.default(buf672, arg379_1, 128, arg380_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg379_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg380_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf674 = buf673
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf673
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_46], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf676 = torch.ops.aten._weight_int4pack_mm.default(buf675, arg382_1, 128, arg383_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg382_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg383_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf677 = buf676
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf676
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf678 = reinterpret_tensor(buf671, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf671  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_220], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf678, arg378_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg378_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf679 = reinterpret_tensor(buf674, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf674  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_220], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf679, arg381_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg381_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf680 = reinterpret_tensor(buf677, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf677  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_220], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf680, arg384_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg384_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_220], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf682 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf678, buf679, buf680, reinterpret_tensor(buf681, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf678
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf679
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf680
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf683 = buf682[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf682
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf687 = buf669; del buf669  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_223], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf683, buf687, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_223], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf688 = torch.ops.aten._weight_int4pack_mm.default(buf687, arg385_1, 128, arg386_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg385_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg386_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf689 = buf688
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf688
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf693 = reinterpret_tensor(buf683, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf683  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_94, hidden_states_225, hidden_states_226, mul_50, norm_hidden_states_46, norm_hidden_states_47], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf689, arg387_1, buf668, arg363_1, buf9, arg14_1, buf693, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf694 = buf687; del buf687  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_227], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf693, buf694, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_227], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf695 = torch.ops.aten._weight_int4pack_mm.default(buf694, arg388_1, 128, arg389_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg388_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg389_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf696 = buf695
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf695
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf697 = buf640; del buf640  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_230], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf696, arg390_1, buf697, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg390_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf696
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_230], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf698 = torch.ops.aten._weight_int4pack_mm.default(buf697, arg391_1, 128, arg392_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg391_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg392_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf699 = buf698
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf698
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf700 = reinterpret_tensor(buf699, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf699  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf704 = buf693; del buf693  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_98, ff_output_11, hidden_states_225, hidden_states_226, hidden_states_231, mul_52, norm_hidden_states_48, norm_hidden_states_49], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf700, arg363_1, buf9, arg14_1, arg393_1, buf689, arg387_1, buf668, arg394_1, buf704, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg363_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg387_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg393_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf668
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf689
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf705 = buf694; del buf694  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf708 = buf651; del buf651  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf711 = buf648; del buf648  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_48, query_48, value_48], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf704, buf705, buf708, buf711, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf704
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_48], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf706 = torch.ops.aten._weight_int4pack_mm.default(buf705, arg395_1, 128, arg396_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg395_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg396_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf707 = buf706
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf706
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_48], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf709 = torch.ops.aten._weight_int4pack_mm.default(buf708, arg398_1, 128, arg399_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg398_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg399_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf710 = buf709
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf709
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_48], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf712 = torch.ops.aten._weight_int4pack_mm.default(buf711, arg401_1, 128, arg402_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg401_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg402_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf713 = buf712
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf712
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf714 = reinterpret_tensor(buf707, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf707  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_232], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf714, arg397_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg397_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf715 = reinterpret_tensor(buf710, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf710  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_232], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf715, arg400_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg400_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf716 = reinterpret_tensor(buf713, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf713  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_232], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf716, arg403_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg403_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_232], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf717 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf714, buf715, buf716, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf714
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf715
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf716
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf718 = buf717[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf717
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf722 = buf711; del buf711  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_235], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf718, buf722, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf718
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_235], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf723 = torch.ops.aten._weight_int4pack_mm.default(buf722, arg404_1, 128, arg405_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg404_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg405_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf724 = buf723
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf723
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf725 = reinterpret_tensor(buf724, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf724  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_12, hidden_states_237, hidden_states_238], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf725, arg394_1, buf9, arg14_1, arg406_1, buf700, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg406_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf700
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf726 = buf722; del buf722  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_50], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf725, buf726, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_50], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf727 = torch.ops.aten._weight_int4pack_mm.default(buf726, arg407_1, 128, arg408_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg407_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg408_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf728 = buf727
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf727
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf729 = buf675; del buf675  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf732 = buf672; del buf672  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf786 = buf618; del buf618  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf789 = buf615; del buf615  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_50, key_54, value_50, value_54], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf729, buf732, buf786, buf789, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_50], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf730 = torch.ops.aten._weight_int4pack_mm.default(buf729, arg410_1, 128, arg411_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg410_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg411_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf731 = buf730
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf730
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_50], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf733 = torch.ops.aten._weight_int4pack_mm.default(buf732, arg413_1, 128, arg414_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg413_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg414_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf734 = buf733
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf733
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf735 = reinterpret_tensor(buf728, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf728  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_239], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf735, arg409_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg409_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf736 = reinterpret_tensor(buf731, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf731  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_239], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf736, arg412_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg412_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf737 = reinterpret_tensor(buf734, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf734  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_239], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf737, arg415_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg415_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf738 = buf681; del buf681  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf795 = buf624; del buf624  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_239, hidden_states_258], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf738, buf795, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_239], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf739 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf735, buf736, buf737, reinterpret_tensor(buf738, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf735
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf736
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf737
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf740 = buf739[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf739
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf744 = buf726; del buf726  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_242], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf740, buf744, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_242], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf745 = torch.ops.aten._weight_int4pack_mm.default(buf744, arg416_1, 128, arg417_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg416_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg417_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf746 = buf745
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf745
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf750 = reinterpret_tensor(buf740, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf740  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_102, hidden_states_244, hidden_states_245, mul_54, norm_hidden_states_50, norm_hidden_states_51], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf746, arg418_1, buf725, arg394_1, buf9, arg14_1, buf750, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf751 = buf744; del buf744  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_246], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf750, buf751, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_246], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf752 = torch.ops.aten._weight_int4pack_mm.default(buf751, arg419_1, 128, arg420_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg419_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg420_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf753 = buf752
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf752
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf754 = buf697; del buf697  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_249], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf753, arg421_1, buf754, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg421_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf753
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_249], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf755 = torch.ops.aten._weight_int4pack_mm.default(buf754, arg422_1, 128, arg423_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg422_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg423_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf756 = buf755
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf755
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf757 = reinterpret_tensor(buf756, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf756  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf761 = buf750; del buf750  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_106, ff_output_12, hidden_states_244, hidden_states_245, hidden_states_250, mul_56, norm_hidden_states_52, norm_hidden_states_53], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf757, arg394_1, buf9, arg14_1, arg424_1, buf746, arg418_1, buf725, arg425_1, buf761, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg394_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg418_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg424_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf725
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf746
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf762 = buf751; del buf751  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf765 = buf708; del buf708  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf768 = buf705; del buf705  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_52, query_52, value_52], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf761, buf762, buf765, buf768, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf761
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_52], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf763 = torch.ops.aten._weight_int4pack_mm.default(buf762, arg426_1, 128, arg427_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg426_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg427_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf764 = buf763
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf763
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_52], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf766 = torch.ops.aten._weight_int4pack_mm.default(buf765, arg429_1, 128, arg430_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg429_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg430_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf767 = buf766
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf766
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_52], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf769 = torch.ops.aten._weight_int4pack_mm.default(buf768, arg432_1, 128, arg433_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg432_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg433_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf770 = buf769
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf769
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf771 = reinterpret_tensor(buf764, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf764  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_251], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf771, arg428_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg428_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf772 = reinterpret_tensor(buf767, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf767  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_251], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf772, arg431_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg431_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf773 = reinterpret_tensor(buf770, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf770  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_251], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf773, arg434_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg434_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_251], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf774 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf771, buf772, buf773, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf771
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf772
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf773
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf775 = buf774[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf774
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf779 = buf768; del buf768  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_254], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf775, buf779, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf775
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_254], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf780 = torch.ops.aten._weight_int4pack_mm.default(buf779, arg435_1, 128, arg436_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg435_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg436_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf781 = buf780
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf780
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf782 = reinterpret_tensor(buf781, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf781  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_13, hidden_states_256, hidden_states_257], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf782, arg425_1, buf9, arg14_1, arg437_1, buf757, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg437_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf757
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf783 = buf779; del buf779  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_54], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf782, buf783, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_54], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf784 = torch.ops.aten._weight_int4pack_mm.default(buf783, arg438_1, 128, arg439_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg438_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg439_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf785 = buf784
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf784
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_54], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf787 = torch.ops.aten._weight_int4pack_mm.default(buf786, arg441_1, 128, arg442_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg441_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg442_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf788 = buf787
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf787
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_54], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf790 = torch.ops.aten._weight_int4pack_mm.default(buf789, arg444_1, 128, arg445_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg444_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg445_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf791 = buf790
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf790
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf792 = reinterpret_tensor(buf785, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf785  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_258], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf792, arg440_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg440_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf793 = reinterpret_tensor(buf788, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf788  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_258], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf793, arg443_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg443_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf794 = reinterpret_tensor(buf791, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf791  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_258], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf794, arg446_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg446_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_258], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf796 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf792, buf793, buf794, reinterpret_tensor(buf795, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf792
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf793
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf794
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf797 = buf796[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf796
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf801 = buf783; del buf783  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_261], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf797, buf801, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_261], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf802 = torch.ops.aten._weight_int4pack_mm.default(buf801, arg447_1, 128, arg448_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg447_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg448_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf803 = buf802
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf802
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf807 = reinterpret_tensor(buf797, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf797  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_110, hidden_states_263, hidden_states_264, mul_58, norm_hidden_states_54, norm_hidden_states_55], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf803, arg449_1, buf782, arg425_1, buf9, arg14_1, buf807, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf808 = buf801; del buf801  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_265], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf807, buf808, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_265], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf809 = torch.ops.aten._weight_int4pack_mm.default(buf808, arg450_1, 128, arg451_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg450_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg451_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf810 = buf809
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf809
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf811 = buf754; del buf754  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_268], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf810, arg452_1, buf811, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg452_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf810
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_268], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf812 = torch.ops.aten._weight_int4pack_mm.default(buf811, arg453_1, 128, arg454_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg453_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg454_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf813 = buf812
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf812
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf814 = reinterpret_tensor(buf813, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf813  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf818 = buf807; del buf807  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_114, ff_output_13, hidden_states_263, hidden_states_264, hidden_states_269, mul_60, norm_hidden_states_56, norm_hidden_states_57], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf814, arg425_1, buf9, arg14_1, arg455_1, buf803, arg449_1, buf782, arg456_1, buf818, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg425_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg449_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg455_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf782
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf803
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf819 = buf808; del buf808  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf822 = buf765; del buf765  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf825 = buf762; del buf762  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_56, query_56, value_56], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf818, buf819, buf822, buf825, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf818
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_56], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf820 = torch.ops.aten._weight_int4pack_mm.default(buf819, arg457_1, 128, arg458_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg457_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg458_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf821 = buf820
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf820
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_56], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf823 = torch.ops.aten._weight_int4pack_mm.default(buf822, arg460_1, 128, arg461_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg460_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg461_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf824 = buf823
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf823
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_56], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf826 = torch.ops.aten._weight_int4pack_mm.default(buf825, arg463_1, 128, arg464_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg463_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg464_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf827 = buf826
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf826
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf828 = reinterpret_tensor(buf821, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf821  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_270], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf828, arg459_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg459_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf829 = reinterpret_tensor(buf824, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf824  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_270], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf829, arg462_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg462_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf830 = reinterpret_tensor(buf827, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf827  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_270], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf830, arg465_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg465_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_270], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf831 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf828, buf829, buf830, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf828
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf829
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf830
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf832 = buf831[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf831
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf836 = buf825; del buf825  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_273], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf832, buf836, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf832
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_273], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf837 = torch.ops.aten._weight_int4pack_mm.default(buf836, arg466_1, 128, arg467_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg466_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg467_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf838 = buf837
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf837
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf839 = reinterpret_tensor(buf838, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf838  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_14, hidden_states_275, hidden_states_276], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf839, arg456_1, buf9, arg14_1, arg468_1, buf814, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg468_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf814
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf840 = buf836; del buf836  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_58], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf839, buf840, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_58], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf841 = torch.ops.aten._weight_int4pack_mm.default(buf840, arg469_1, 128, arg470_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg469_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg470_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf842 = buf841
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf841
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf843 = buf789; del buf789  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf846 = buf786; del buf786  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf900 = buf732; del buf732  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf903 = buf729; del buf729  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_58, key_62, value_58, value_62], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf843, buf846, buf900, buf903, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_58], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf844 = torch.ops.aten._weight_int4pack_mm.default(buf843, arg472_1, 128, arg473_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg472_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg473_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf845 = buf844
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf844
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_58], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf847 = torch.ops.aten._weight_int4pack_mm.default(buf846, arg475_1, 128, arg476_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg475_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg476_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf848 = buf847
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf847
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf849 = reinterpret_tensor(buf842, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf842  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_277], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf849, arg471_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg471_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf850 = reinterpret_tensor(buf845, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf845  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_277], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf850, arg474_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg474_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf851 = reinterpret_tensor(buf848, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf848  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_277], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf851, arg477_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg477_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf852 = buf795; del buf795  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf909 = buf738; del buf738  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_277, hidden_states_296], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf852, buf909, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_277], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf853 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf849, buf850, buf851, reinterpret_tensor(buf852, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf849
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf850
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf851
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf854 = buf853[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf853
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf858 = buf840; del buf840  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_280], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf854, buf858, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_280], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf859 = torch.ops.aten._weight_int4pack_mm.default(buf858, arg478_1, 128, arg479_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg478_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg479_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf860 = buf859
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf859
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf864 = reinterpret_tensor(buf854, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf854  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_118, hidden_states_282, hidden_states_283, mul_62, norm_hidden_states_58, norm_hidden_states_59], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf860, arg480_1, buf839, arg456_1, buf9, arg14_1, buf864, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf865 = buf858; del buf858  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_284], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf864, buf865, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_284], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf866 = torch.ops.aten._weight_int4pack_mm.default(buf865, arg481_1, 128, arg482_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg481_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg482_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf867 = buf866
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf866
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf868 = buf811; del buf811  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_287], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf867, arg483_1, buf868, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg483_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf867
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_287], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf869 = torch.ops.aten._weight_int4pack_mm.default(buf868, arg484_1, 128, arg485_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg484_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg485_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf870 = buf869
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf869
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf871 = reinterpret_tensor(buf870, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf870  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf875 = buf864; del buf864  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_122, ff_output_14, hidden_states_282, hidden_states_283, hidden_states_288, mul_64, norm_hidden_states_60, norm_hidden_states_61], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf871, arg456_1, buf9, arg14_1, arg486_1, buf860, arg480_1, buf839, arg487_1, buf875, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg456_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg480_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg486_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf839
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf860
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf876 = buf865; del buf865  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf879 = buf822; del buf822  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf882 = buf819; del buf819  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_60, query_60, value_60], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf875, buf876, buf879, buf882, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf875
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_60], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf877 = torch.ops.aten._weight_int4pack_mm.default(buf876, arg488_1, 128, arg489_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg488_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg489_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf878 = buf877
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf877
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_60], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf880 = torch.ops.aten._weight_int4pack_mm.default(buf879, arg491_1, 128, arg492_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg491_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg492_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf881 = buf880
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf880
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_60], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf883 = torch.ops.aten._weight_int4pack_mm.default(buf882, arg494_1, 128, arg495_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg494_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg495_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf884 = buf883
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf883
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf885 = reinterpret_tensor(buf878, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf878  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_289], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf885, arg490_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg490_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf886 = reinterpret_tensor(buf881, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf881  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_289], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf886, arg493_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg493_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf887 = reinterpret_tensor(buf884, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf884  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_289], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf887, arg496_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg496_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_289], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf888 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf885, buf886, buf887, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf885
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf886
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf887
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf889 = buf888[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf888
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf893 = buf882; del buf882  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_292], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf889, buf893, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf889
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_292], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf894 = torch.ops.aten._weight_int4pack_mm.default(buf893, arg497_1, 128, arg498_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg497_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg498_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf895 = buf894
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf894
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf896 = reinterpret_tensor(buf895, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf895  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_15, hidden_states_294, hidden_states_295], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf896, arg487_1, buf9, arg14_1, arg499_1, buf871, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg499_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf871
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf897 = buf893; del buf893  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_62], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf896, buf897, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_62], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf898 = torch.ops.aten._weight_int4pack_mm.default(buf897, arg500_1, 128, arg501_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg500_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg501_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf899 = buf898
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf898
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_62], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf901 = torch.ops.aten._weight_int4pack_mm.default(buf900, arg503_1, 128, arg504_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg503_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg504_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf902 = buf901
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf901
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_62], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf904 = torch.ops.aten._weight_int4pack_mm.default(buf903, arg506_1, 128, arg507_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg506_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg507_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf905 = buf904
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf904
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf906 = reinterpret_tensor(buf899, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf899  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_296], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf906, arg502_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg502_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf907 = reinterpret_tensor(buf902, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf902  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_296], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf907, arg505_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg505_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf908 = reinterpret_tensor(buf905, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf905  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_296], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf908, arg508_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg508_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_296], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf910 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf906, buf907, buf908, reinterpret_tensor(buf909, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf906
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf907
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf908
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf911 = buf910[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf910
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf915 = buf897; del buf897  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_299], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf911, buf915, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_299], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf916 = torch.ops.aten._weight_int4pack_mm.default(buf915, arg509_1, 128, arg510_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg509_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg510_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf917 = buf916
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf916
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf921 = reinterpret_tensor(buf911, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf911  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_126, hidden_states_301, hidden_states_302, mul_66, norm_hidden_states_62, norm_hidden_states_63], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf917, arg511_1, buf896, arg487_1, buf9, arg14_1, buf921, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf922 = buf915; del buf915  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_303], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf921, buf922, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_303], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf923 = torch.ops.aten._weight_int4pack_mm.default(buf922, arg512_1, 128, arg513_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg512_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg513_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf924 = buf923
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf923
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf925 = buf868; del buf868  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_306], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf924, arg514_1, buf925, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg514_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf924
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_306], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf926 = torch.ops.aten._weight_int4pack_mm.default(buf925, arg515_1, 128, arg516_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg515_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg516_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf927 = buf926
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf926
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf928 = reinterpret_tensor(buf927, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf927  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf932 = buf921; del buf921  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_130, ff_output_15, hidden_states_301, hidden_states_302, hidden_states_307, mul_68, norm_hidden_states_64, norm_hidden_states_65], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf928, arg487_1, buf9, arg14_1, arg517_1, buf917, arg511_1, buf896, arg518_1, buf932, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg487_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg511_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg517_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf896
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf917
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf933 = buf922; del buf922  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf936 = buf879; del buf879  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf939 = buf876; del buf876  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_64, query_64, value_64], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf932, buf933, buf936, buf939, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf932
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_64], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf934 = torch.ops.aten._weight_int4pack_mm.default(buf933, arg519_1, 128, arg520_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg519_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg520_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf935 = buf934
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf934
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_64], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf937 = torch.ops.aten._weight_int4pack_mm.default(buf936, arg522_1, 128, arg523_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg522_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg523_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf938 = buf937
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf937
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_64], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf940 = torch.ops.aten._weight_int4pack_mm.default(buf939, arg525_1, 128, arg526_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg525_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg526_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf941 = buf940
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf940
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf942 = reinterpret_tensor(buf935, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf935  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_308], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf942, arg521_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg521_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf943 = reinterpret_tensor(buf938, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf938  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_308], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf943, arg524_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg524_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf944 = reinterpret_tensor(buf941, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf941  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_308], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf944, arg527_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg527_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_308], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf945 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf942, buf943, buf944, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf942
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf943
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf944
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf946 = buf945[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf945
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf950 = buf939; del buf939  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_311], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf946, buf950, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf946
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_311], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf951 = torch.ops.aten._weight_int4pack_mm.default(buf950, arg528_1, 128, arg529_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg528_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg529_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf952 = buf951
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf951
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf953 = reinterpret_tensor(buf952, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf952  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_16, hidden_states_313, hidden_states_314], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf953, arg518_1, buf9, arg14_1, arg530_1, buf928, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg530_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf928
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf954 = buf950; del buf950  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_66], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf953, buf954, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_66], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf955 = torch.ops.aten._weight_int4pack_mm.default(buf954, arg531_1, 128, arg532_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg531_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg532_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf956 = buf955
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf955
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf957 = buf903; del buf903  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf960 = buf900; del buf900  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1014 = buf846; del buf846  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1017 = buf843; del buf843  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_66, key_70, value_66, value_70], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf957, buf960, buf1014, buf1017, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_66], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf958 = torch.ops.aten._weight_int4pack_mm.default(buf957, arg534_1, 128, arg535_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg534_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg535_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf959 = buf958
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf958
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_66], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf961 = torch.ops.aten._weight_int4pack_mm.default(buf960, arg537_1, 128, arg538_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg537_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg538_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf962 = buf961
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf961
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf963 = reinterpret_tensor(buf956, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf956  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_315], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf963, arg533_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg533_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf964 = reinterpret_tensor(buf959, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf959  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_315], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf964, arg536_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg536_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf965 = reinterpret_tensor(buf962, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf962  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_315], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf965, arg539_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg539_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf966 = buf909; del buf909  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1023 = buf852; del buf852  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_315, hidden_states_334], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf966, buf1023, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_315], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf967 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf963, buf964, buf965, reinterpret_tensor(buf966, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf963
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf964
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf965
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf968 = buf967[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf967
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf972 = buf954; del buf954  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_318], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf968, buf972, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_318], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf973 = torch.ops.aten._weight_int4pack_mm.default(buf972, arg540_1, 128, arg541_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg540_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg541_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf974 = buf973
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf973
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf978 = reinterpret_tensor(buf968, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf968  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_134, hidden_states_320, hidden_states_321, mul_70, norm_hidden_states_66, norm_hidden_states_67], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf974, arg542_1, buf953, arg518_1, buf9, arg14_1, buf978, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf979 = buf972; del buf972  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_322], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf978, buf979, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_322], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf980 = torch.ops.aten._weight_int4pack_mm.default(buf979, arg543_1, 128, arg544_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg543_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg544_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf981 = buf980
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf980
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf982 = buf925; del buf925  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_325], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf981, arg545_1, buf982, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg545_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf981
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_325], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf983 = torch.ops.aten._weight_int4pack_mm.default(buf982, arg546_1, 128, arg547_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg546_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg547_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf984 = buf983
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf983
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf985 = reinterpret_tensor(buf984, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf984  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf989 = buf978; del buf978  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_138, ff_output_16, hidden_states_320, hidden_states_321, hidden_states_326, mul_72, norm_hidden_states_68, norm_hidden_states_69], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf985, arg518_1, buf9, arg14_1, arg548_1, buf974, arg542_1, buf953, arg549_1, buf989, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg518_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg542_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg548_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf953
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf974
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf990 = buf979; del buf979  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf993 = buf936; del buf936  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf996 = buf933; del buf933  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_68, query_68, value_68], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf989, buf990, buf993, buf996, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf989
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_68], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf991 = torch.ops.aten._weight_int4pack_mm.default(buf990, arg550_1, 128, arg551_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg550_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg551_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf992 = buf991
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf991
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_68], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf994 = torch.ops.aten._weight_int4pack_mm.default(buf993, arg553_1, 128, arg554_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg553_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg554_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf995 = buf994
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf994
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_68], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf997 = torch.ops.aten._weight_int4pack_mm.default(buf996, arg556_1, 128, arg557_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg556_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg557_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf998 = buf997
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf997
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf999 = reinterpret_tensor(buf992, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf992  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_327], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf999, arg552_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg552_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1000 = reinterpret_tensor(buf995, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf995  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_327], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1000, arg555_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg555_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1001 = reinterpret_tensor(buf998, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf998  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_327], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1001, arg558_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg558_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_327], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1002 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf999, buf1000, buf1001, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1000
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1001
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf999
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1003 = buf1002[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1002
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1007 = buf996; del buf996  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_330], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf1003, buf1007, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1003
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_330], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1008 = torch.ops.aten._weight_int4pack_mm.default(buf1007, arg559_1, 128, arg560_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg559_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg560_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1009 = buf1008
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1008
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1010 = reinterpret_tensor(buf1009, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1009  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_17, hidden_states_332, hidden_states_333], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf1010, arg549_1, buf9, arg14_1, arg561_1, buf985, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg561_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf985
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1011 = buf1007; del buf1007  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_70], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1010, buf1011, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_70], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1012 = torch.ops.aten._weight_int4pack_mm.default(buf1011, arg562_1, 128, arg563_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg562_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg563_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1013 = buf1012
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1012
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_70], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1015 = torch.ops.aten._weight_int4pack_mm.default(buf1014, arg565_1, 128, arg566_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg565_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg566_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1016 = buf1015
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1015
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_70], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1018 = torch.ops.aten._weight_int4pack_mm.default(buf1017, arg568_1, 128, arg569_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg568_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg569_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1019 = buf1018
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1018
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1020 = reinterpret_tensor(buf1013, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1013  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_334], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1020, arg564_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg564_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1021 = reinterpret_tensor(buf1016, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1016  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_334], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1021, arg567_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg567_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1022 = reinterpret_tensor(buf1019, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1019  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_334], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1022, arg570_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg570_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_334], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1024 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1020, buf1021, buf1022, reinterpret_tensor(buf1023, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1020
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1021
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1022
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1025 = buf1024[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1024
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1029 = buf1011; del buf1011  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_337], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1025, buf1029, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_337], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1030 = torch.ops.aten._weight_int4pack_mm.default(buf1029, arg571_1, 128, arg572_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg571_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg572_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1031 = buf1030
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1030
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1035 = reinterpret_tensor(buf1025, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1025  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_142, hidden_states_339, hidden_states_340, mul_74, norm_hidden_states_70, norm_hidden_states_71], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1031, arg573_1, buf1010, arg549_1, buf9, arg14_1, buf1035, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1036 = buf1029; del buf1029  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_341], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1035, buf1036, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_341], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1037 = torch.ops.aten._weight_int4pack_mm.default(buf1036, arg574_1, 128, arg575_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg574_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg575_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1038 = buf1037
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1037
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1039 = buf982; del buf982  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_344], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf1038, arg576_1, buf1039, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg576_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1038
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_344], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1040 = torch.ops.aten._weight_int4pack_mm.default(buf1039, arg577_1, 128, arg578_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg577_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg578_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1041 = buf1040
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1040
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1042 = reinterpret_tensor(buf1041, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1041  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1046 = buf1035; del buf1035  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_146, ff_output_17, hidden_states_339, hidden_states_340, hidden_states_345, mul_76, norm_hidden_states_72, norm_hidden_states_73], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1042, arg549_1, buf9, arg14_1, arg579_1, buf1031, arg573_1, buf1010, arg580_1, buf1046, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg549_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg573_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg579_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1010
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1031
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1047 = buf1036; del buf1036  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1050 = buf993; del buf993  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1053 = buf990; del buf990  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_72, query_72, value_72], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf1046, buf1047, buf1050, buf1053, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1046
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_72], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1048 = torch.ops.aten._weight_int4pack_mm.default(buf1047, arg581_1, 128, arg582_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg581_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg582_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1049 = buf1048
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1048
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_72], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1051 = torch.ops.aten._weight_int4pack_mm.default(buf1050, arg584_1, 128, arg585_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg584_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg585_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1052 = buf1051
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1051
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_72], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1054 = torch.ops.aten._weight_int4pack_mm.default(buf1053, arg587_1, 128, arg588_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg587_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg588_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1055 = buf1054
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1054
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1056 = reinterpret_tensor(buf1049, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1049  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_346], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1056, arg583_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg583_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1057 = reinterpret_tensor(buf1052, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1052  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_346], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1057, arg586_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg586_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1058 = reinterpret_tensor(buf1055, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1055  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_346], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1058, arg589_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg589_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_346], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1059 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1056, buf1057, buf1058, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1056
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1057
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1058
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1060 = buf1059[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1059
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1064 = buf1053; del buf1053  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_349], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf1060, buf1064, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1060
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_349], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1065 = torch.ops.aten._weight_int4pack_mm.default(buf1064, arg590_1, 128, arg591_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg590_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg591_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1066 = buf1065
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1065
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1067 = reinterpret_tensor(buf1066, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1066  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_18, hidden_states_351, hidden_states_352], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf1067, arg580_1, buf9, arg14_1, arg592_1, buf1042, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg592_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1042
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1068 = buf1064; del buf1064  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_74], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1067, buf1068, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_74], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1069 = torch.ops.aten._weight_int4pack_mm.default(buf1068, arg593_1, 128, arg594_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg593_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg594_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1070 = buf1069
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1069
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1071 = buf1017; del buf1017  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1074 = buf1014; del buf1014  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1128 = buf960; del buf960  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1131 = buf957; del buf957  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_74, key_78, value_74, value_78], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf1071, buf1074, buf1128, buf1131, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_74], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1072 = torch.ops.aten._weight_int4pack_mm.default(buf1071, arg596_1, 128, arg597_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg596_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg597_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1073 = buf1072
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1072
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_74], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1075 = torch.ops.aten._weight_int4pack_mm.default(buf1074, arg599_1, 128, arg600_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg599_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg600_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1076 = buf1075
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1075
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1077 = reinterpret_tensor(buf1070, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1070  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_353], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1077, arg595_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg595_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1078 = reinterpret_tensor(buf1073, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1073  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_353], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1078, arg598_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg598_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1079 = reinterpret_tensor(buf1076, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1076  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_353], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1079, arg601_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg601_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1080 = buf1023; del buf1023  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1137 = buf966; del buf966  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_353, hidden_states_372], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf1080, buf1137, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_353], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1081 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1077, buf1078, buf1079, reinterpret_tensor(buf1080, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1077
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1078
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1079
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1082 = buf1081[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1081
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1086 = buf1068; del buf1068  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_356], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1082, buf1086, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_356], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1087 = torch.ops.aten._weight_int4pack_mm.default(buf1086, arg602_1, 128, arg603_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg602_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg603_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1088 = buf1087
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1087
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1092 = reinterpret_tensor(buf1082, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1082  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_150, hidden_states_358, hidden_states_359, mul_78, norm_hidden_states_74, norm_hidden_states_75], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1088, arg604_1, buf1067, arg580_1, buf9, arg14_1, buf1092, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1093 = buf1086; del buf1086  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_360], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1092, buf1093, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_360], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1094 = torch.ops.aten._weight_int4pack_mm.default(buf1093, arg605_1, 128, arg606_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg605_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg606_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1095 = buf1094
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1094
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1096 = buf1039; del buf1039  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_363], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf1095, arg607_1, buf1096, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg607_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1095
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_363], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1097 = torch.ops.aten._weight_int4pack_mm.default(buf1096, arg608_1, 128, arg609_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg608_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg609_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1098 = buf1097
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1097
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1099 = reinterpret_tensor(buf1098, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1098  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1103 = buf1092; del buf1092  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_154, ff_output_18, hidden_states_358, hidden_states_359, hidden_states_364, mul_80, norm_hidden_states_76, norm_hidden_states_77], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1099, arg580_1, buf9, arg14_1, arg610_1, buf1088, arg604_1, buf1067, arg611_1, buf1103, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg580_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg604_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg610_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1067
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1088
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1104 = buf1093; del buf1093  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1107 = buf1050; del buf1050  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1110 = buf1047; del buf1047  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_76, query_76, value_76], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf1103, buf1104, buf1107, buf1110, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1103
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_76], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1105 = torch.ops.aten._weight_int4pack_mm.default(buf1104, arg612_1, 128, arg613_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg612_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg613_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1106 = buf1105
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1105
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_76], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1108 = torch.ops.aten._weight_int4pack_mm.default(buf1107, arg615_1, 128, arg616_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg615_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg616_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1109 = buf1108
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1108
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_76], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1111 = torch.ops.aten._weight_int4pack_mm.default(buf1110, arg618_1, 128, arg619_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg618_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg619_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1112 = buf1111
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1111
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1113 = reinterpret_tensor(buf1106, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1106  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_365], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1113, arg614_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg614_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1114 = reinterpret_tensor(buf1109, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1109  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_365], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1114, arg617_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg617_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1115 = reinterpret_tensor(buf1112, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1112  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_365], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1115, arg620_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg620_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_365], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1116 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1113, buf1114, buf1115, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1113
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1114
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1115
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1117 = buf1116[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1116
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1121 = buf1110; del buf1110  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_368], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf1117, buf1121, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1117
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_368], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1122 = torch.ops.aten._weight_int4pack_mm.default(buf1121, arg621_1, 128, arg622_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg621_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg622_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1123 = buf1122
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1122
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1124 = reinterpret_tensor(buf1123, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1123  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_19, hidden_states_370, hidden_states_371], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf1124, arg611_1, buf9, arg14_1, arg623_1, buf1099, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg623_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1099
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1125 = buf1121; del buf1121  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_78], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1124, buf1125, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_78], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1126 = torch.ops.aten._weight_int4pack_mm.default(buf1125, arg624_1, 128, arg625_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg624_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg625_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1127 = buf1126
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1126
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_78], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1129 = torch.ops.aten._weight_int4pack_mm.default(buf1128, arg627_1, 128, arg628_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg627_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg628_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1130 = buf1129
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1129
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_78], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1132 = torch.ops.aten._weight_int4pack_mm.default(buf1131, arg630_1, 128, arg631_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg630_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg631_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1133 = buf1132
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1132
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1134 = reinterpret_tensor(buf1127, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1127  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_372], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1134, arg626_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg626_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1135 = reinterpret_tensor(buf1130, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1130  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_372], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1135, arg629_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg629_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1136 = reinterpret_tensor(buf1133, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1133  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_372], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1136, arg632_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg632_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_372], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1138 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1134, buf1135, buf1136, reinterpret_tensor(buf1137, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1134
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1135
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1136
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1139 = buf1138[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1138
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1143 = buf1125; del buf1125  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_375], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1139, buf1143, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_375], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1144 = torch.ops.aten._weight_int4pack_mm.default(buf1143, arg633_1, 128, arg634_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg633_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg634_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1145 = buf1144
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1144
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1149 = reinterpret_tensor(buf1139, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1139  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_158, hidden_states_377, hidden_states_378, mul_82, norm_hidden_states_78, norm_hidden_states_79], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1145, arg635_1, buf1124, arg611_1, buf9, arg14_1, buf1149, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1150 = buf1143; del buf1143  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_379], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1149, buf1150, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_379], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1151 = torch.ops.aten._weight_int4pack_mm.default(buf1150, arg636_1, 128, arg637_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg636_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg637_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1152 = buf1151
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1151
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1153 = buf1096; del buf1096  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_382], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf1152, arg638_1, buf1153, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg638_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1152
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_382], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1154 = torch.ops.aten._weight_int4pack_mm.default(buf1153, arg639_1, 128, arg640_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg639_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg640_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1155 = buf1154
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1154
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1156 = reinterpret_tensor(buf1155, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1155  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1160 = buf1149; del buf1149  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_162, ff_output_19, hidden_states_377, hidden_states_378, hidden_states_383, mul_84, norm_hidden_states_80, norm_hidden_states_81], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1156, arg611_1, buf9, arg14_1, arg641_1, buf1145, arg635_1, buf1124, arg642_1, buf1160, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg611_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg635_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg641_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1124
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1145
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1161 = buf1150; del buf1150  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1164 = buf1107; del buf1107  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1167 = buf1104; del buf1104  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_80, query_80, value_80], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf1160, buf1161, buf1164, buf1167, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1160
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_80], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1162 = torch.ops.aten._weight_int4pack_mm.default(buf1161, arg643_1, 128, arg644_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg643_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg644_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1163 = buf1162
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1162
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_80], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1165 = torch.ops.aten._weight_int4pack_mm.default(buf1164, arg646_1, 128, arg647_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg646_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg647_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1166 = buf1165
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1165
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_80], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1168 = torch.ops.aten._weight_int4pack_mm.default(buf1167, arg649_1, 128, arg650_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg649_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg650_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1169 = buf1168
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1168
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1170 = reinterpret_tensor(buf1163, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1163  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_384], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1170, arg645_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg645_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1171 = reinterpret_tensor(buf1166, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1166  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_384], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1171, arg648_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg648_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1172 = reinterpret_tensor(buf1169, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1169  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_384], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1172, arg651_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg651_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_384], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1173 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1170, buf1171, buf1172, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1170
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1171
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1172
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1174 = buf1173[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1173
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1178 = buf1167; del buf1167  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_387], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf1174, buf1178, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1174
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_387], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1179 = torch.ops.aten._weight_int4pack_mm.default(buf1178, arg652_1, 128, arg653_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg652_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg653_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1180 = buf1179
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1179
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1181 = reinterpret_tensor(buf1180, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1180  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_20, hidden_states_389, hidden_states_390], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf1181, arg642_1, buf9, arg14_1, arg654_1, buf1156, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg654_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1156
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1182 = buf1178; del buf1178  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_82], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1181, buf1182, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_82], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1183 = torch.ops.aten._weight_int4pack_mm.default(buf1182, arg655_1, 128, arg656_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg655_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg656_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1184 = buf1183
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1183
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1185 = buf1131; del buf1131  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1188 = buf1128; del buf1128  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1242 = buf1074; del buf1074  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1245 = buf1071; del buf1071  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_82, key_86, value_82, value_86], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf1185, buf1188, buf1242, buf1245, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_82], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1186 = torch.ops.aten._weight_int4pack_mm.default(buf1185, arg658_1, 128, arg659_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg658_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg659_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1187 = buf1186
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1186
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_82], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1189 = torch.ops.aten._weight_int4pack_mm.default(buf1188, arg661_1, 128, arg662_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg661_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg662_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1190 = buf1189
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1189
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1191 = reinterpret_tensor(buf1184, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1184  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_391], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1191, arg657_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg657_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1192 = reinterpret_tensor(buf1187, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1187  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_391], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1192, arg660_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg660_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1193 = reinterpret_tensor(buf1190, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1190  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_391], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1193, arg663_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg663_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1194 = buf1137; del buf1137  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1251 = buf1080; del buf1080  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_391, hidden_states_410], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf1194, buf1251, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_391], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1195 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1191, buf1192, buf1193, reinterpret_tensor(buf1194, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1191
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1192
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1193
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1196 = buf1195[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1195
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1200 = buf1182; del buf1182  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_394], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1196, buf1200, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_394], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1201 = torch.ops.aten._weight_int4pack_mm.default(buf1200, arg664_1, 128, arg665_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg664_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg665_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1202 = buf1201
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1201
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1206 = reinterpret_tensor(buf1196, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1196  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_166, hidden_states_396, hidden_states_397, mul_86, norm_hidden_states_82, norm_hidden_states_83], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1202, arg666_1, buf1181, arg642_1, buf9, arg14_1, buf1206, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1207 = buf1200; del buf1200  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_398], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1206, buf1207, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_398], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1208 = torch.ops.aten._weight_int4pack_mm.default(buf1207, arg667_1, 128, arg668_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg667_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg668_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1209 = buf1208
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1208
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1210 = buf1153; del buf1153  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_401], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf1209, arg669_1, buf1210, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg669_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1209
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_401], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1211 = torch.ops.aten._weight_int4pack_mm.default(buf1210, arg670_1, 128, arg671_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg670_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg671_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1212 = buf1211
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1211
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1213 = reinterpret_tensor(buf1212, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1212  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1217 = buf1206; del buf1206  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_170, ff_output_20, hidden_states_396, hidden_states_397, hidden_states_402, mul_88, norm_hidden_states_84, norm_hidden_states_85], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1213, arg642_1, buf9, arg14_1, arg672_1, buf1202, arg666_1, buf1181, arg673_1, buf1217, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg642_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg666_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg672_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1181
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1202
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1218 = buf1207; del buf1207  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1221 = buf1164; del buf1164  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1224 = buf1161; del buf1161  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_84, query_84, value_84], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf1217, buf1218, buf1221, buf1224, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1217
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_84], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1219 = torch.ops.aten._weight_int4pack_mm.default(buf1218, arg674_1, 128, arg675_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg674_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg675_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1220 = buf1219
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1219
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_84], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1222 = torch.ops.aten._weight_int4pack_mm.default(buf1221, arg677_1, 128, arg678_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg677_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg678_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1223 = buf1222
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1222
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_84], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1225 = torch.ops.aten._weight_int4pack_mm.default(buf1224, arg680_1, 128, arg681_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg680_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg681_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1226 = buf1225
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1225
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1227 = reinterpret_tensor(buf1220, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1220  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_403], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1227, arg676_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg676_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1228 = reinterpret_tensor(buf1223, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1223  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_403], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1228, arg679_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg679_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1229 = reinterpret_tensor(buf1226, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1226  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_403], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1229, arg682_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg682_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_403], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1230 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1227, buf1228, buf1229, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1227
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1228
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1229
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1231 = buf1230[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1230
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1235 = buf1224; del buf1224  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_406], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf1231, buf1235, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1231
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_406], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1236 = torch.ops.aten._weight_int4pack_mm.default(buf1235, arg683_1, 128, arg684_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg683_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg684_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1237 = buf1236
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1236
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1238 = reinterpret_tensor(buf1237, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1237  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_21, hidden_states_408, hidden_states_409], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf1238, arg673_1, buf9, arg14_1, arg685_1, buf1213, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg685_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1213
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1239 = buf1235; del buf1235  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_86], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1238, buf1239, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_86], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1240 = torch.ops.aten._weight_int4pack_mm.default(buf1239, arg686_1, 128, arg687_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg686_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg687_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1241 = buf1240
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1240
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_86], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1243 = torch.ops.aten._weight_int4pack_mm.default(buf1242, arg689_1, 128, arg690_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg689_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg690_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1244 = buf1243
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1243
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_86], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1246 = torch.ops.aten._weight_int4pack_mm.default(buf1245, arg692_1, 128, arg693_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg692_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg693_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1247 = buf1246
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1246
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1248 = reinterpret_tensor(buf1241, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1241  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_410], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1248, arg688_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg688_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1249 = reinterpret_tensor(buf1244, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1244  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_410], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1249, arg691_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg691_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1250 = reinterpret_tensor(buf1247, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1247  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_410], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1250, arg694_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg694_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_410], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1252 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1248, buf1249, buf1250, reinterpret_tensor(buf1251, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1248
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1249
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1250
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1253 = buf1252[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1252
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1257 = buf1239; del buf1239  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_413], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1253, buf1257, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_413], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1258 = torch.ops.aten._weight_int4pack_mm.default(buf1257, arg695_1, 128, arg696_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg695_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg696_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1259 = buf1258
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1258
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1263 = reinterpret_tensor(buf1253, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1253  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_174, hidden_states_415, hidden_states_416, mul_90, norm_hidden_states_86, norm_hidden_states_87], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1259, arg697_1, buf1238, arg673_1, buf9, arg14_1, buf1263, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1264 = buf1257; del buf1257  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_417], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1263, buf1264, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_417], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1265 = torch.ops.aten._weight_int4pack_mm.default(buf1264, arg698_1, 128, arg699_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg698_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg699_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1266 = buf1265
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1265
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1267 = buf1210; del buf1210  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_420], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf1266, arg700_1, buf1267, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg700_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1266
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_420], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1268 = torch.ops.aten._weight_int4pack_mm.default(buf1267, arg701_1, 128, arg702_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg701_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg702_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1269 = buf1268
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1268
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1270 = reinterpret_tensor(buf1269, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1269  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1274 = buf1263; del buf1263  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_178, ff_output_21, hidden_states_415, hidden_states_416, hidden_states_421, mul_92, norm_hidden_states_88, norm_hidden_states_89], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1270, arg673_1, buf9, arg14_1, arg703_1, buf1259, arg697_1, buf1238, arg704_1, buf1274, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg673_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg697_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg703_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1238
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1259
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1275 = buf1264; del buf1264  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1278 = buf1221; del buf1221  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1281 = buf1218; del buf1218  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_88, query_88, value_88], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf1274, buf1275, buf1278, buf1281, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1274
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_88], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1276 = torch.ops.aten._weight_int4pack_mm.default(buf1275, arg705_1, 128, arg706_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg705_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg706_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1277 = buf1276
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1276
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_88], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1279 = torch.ops.aten._weight_int4pack_mm.default(buf1278, arg708_1, 128, arg709_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg708_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg709_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1280 = buf1279
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1279
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_88], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1282 = torch.ops.aten._weight_int4pack_mm.default(buf1281, arg711_1, 128, arg712_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg711_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg712_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1283 = buf1282
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1282
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1284 = reinterpret_tensor(buf1277, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1277  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_422], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1284, arg707_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg707_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1285 = reinterpret_tensor(buf1280, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1280  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_422], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1285, arg710_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg710_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1286 = reinterpret_tensor(buf1283, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1283  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_422], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1286, arg713_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg713_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_422], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1287 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1284, buf1285, buf1286, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1284
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1285
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1286
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1288 = buf1287[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1287
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1292 = buf1281; del buf1281  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_425], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf1288, buf1292, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1288
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_425], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1293 = torch.ops.aten._weight_int4pack_mm.default(buf1292, arg714_1, 128, arg715_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg714_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg715_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1294 = buf1293
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1293
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1295 = reinterpret_tensor(buf1294, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1294  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_22, hidden_states_427, hidden_states_428], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf1295, arg704_1, buf9, arg14_1, arg716_1, buf1270, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg716_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1270
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1296 = buf1292; del buf1292  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_90], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1295, buf1296, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_90], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1297 = torch.ops.aten._weight_int4pack_mm.default(buf1296, arg717_1, 128, arg718_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg717_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg718_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1298 = buf1297
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1297
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1299 = buf1245; del buf1245  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1302 = buf1242; del buf1242  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1356 = buf1188; del buf1188  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1359 = buf1185; del buf1185  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_90, key_94, value_90, value_94], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf1299, buf1302, buf1356, buf1359, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_90], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1300 = torch.ops.aten._weight_int4pack_mm.default(buf1299, arg720_1, 128, arg721_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg720_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg721_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1301 = buf1300
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1300
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_90], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1303 = torch.ops.aten._weight_int4pack_mm.default(buf1302, arg723_1, 128, arg724_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg723_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg724_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1304 = buf1303
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1303
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1305 = reinterpret_tensor(buf1298, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1298  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_429], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1305, arg719_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg719_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1306 = reinterpret_tensor(buf1301, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1301  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_429], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1306, arg722_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg722_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1307 = reinterpret_tensor(buf1304, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1304  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_429], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1307, arg725_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg725_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1308 = buf1251; del buf1251  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1365 = buf1194; del buf1194  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_429, hidden_states_448], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf1308, buf1365, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_429], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1309 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1305, buf1306, buf1307, reinterpret_tensor(buf1308, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1305
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1306
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1307
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1310 = buf1309[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1309
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1314 = buf1296; del buf1296  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_432], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1310, buf1314, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_432], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1315 = torch.ops.aten._weight_int4pack_mm.default(buf1314, arg726_1, 128, arg727_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg726_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg727_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1316 = buf1315
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1315
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1320 = reinterpret_tensor(buf1310, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1310  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_182, hidden_states_434, hidden_states_435, mul_94, norm_hidden_states_90, norm_hidden_states_91], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1316, arg728_1, buf1295, arg704_1, buf9, arg14_1, buf1320, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1321 = buf1314; del buf1314  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_436], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1320, buf1321, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_436], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1322 = torch.ops.aten._weight_int4pack_mm.default(buf1321, arg729_1, 128, arg730_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg729_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg730_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1323 = buf1322
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1322
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1324 = buf1267; del buf1267  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_439], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf1323, arg731_1, buf1324, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg731_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1323
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_439], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1325 = torch.ops.aten._weight_int4pack_mm.default(buf1324, arg732_1, 128, arg733_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg732_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg733_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1326 = buf1325
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1325
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1327 = reinterpret_tensor(buf1326, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1326  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1331 = buf1320; del buf1320  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_186, ff_output_22, hidden_states_434, hidden_states_435, hidden_states_440, mul_96, norm_hidden_states_92, norm_hidden_states_93], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1327, arg704_1, buf9, arg14_1, arg734_1, buf1316, arg728_1, buf1295, arg735_1, buf1331, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg704_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg728_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg734_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1295
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1316
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1332 = buf1321; del buf1321  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1335 = buf1278; del buf1278  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1338 = buf1275; del buf1275  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_92, query_92, value_92], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf1331, buf1332, buf1335, buf1338, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1331
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_92], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1333 = torch.ops.aten._weight_int4pack_mm.default(buf1332, arg736_1, 128, arg737_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg736_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg737_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1334 = buf1333
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1333
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_92], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1336 = torch.ops.aten._weight_int4pack_mm.default(buf1335, arg739_1, 128, arg740_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg739_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg740_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1337 = buf1336
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1336
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_92], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1339 = torch.ops.aten._weight_int4pack_mm.default(buf1338, arg742_1, 128, arg743_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg742_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg743_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1340 = buf1339
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1339
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1341 = reinterpret_tensor(buf1334, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1334  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_441], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1341, arg738_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg738_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1342 = reinterpret_tensor(buf1337, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1337  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_441], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1342, arg741_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg741_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1343 = reinterpret_tensor(buf1340, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1340  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_441], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1343, arg744_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg744_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_441], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1344 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1341, buf1342, buf1343, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1341
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1342
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1343
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1345 = buf1344[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1344
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1349 = buf1338; del buf1338  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_444], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf1345, buf1349, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1345
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_444], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1350 = torch.ops.aten._weight_int4pack_mm.default(buf1349, arg745_1, 128, arg746_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg745_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg746_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1351 = buf1350
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1350
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1352 = reinterpret_tensor(buf1351, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1351  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_23, hidden_states_446, hidden_states_447], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf1352, arg735_1, buf9, arg14_1, arg747_1, buf1327, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg747_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1327
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1353 = buf1349; del buf1349  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_94], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1352, buf1353, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_94], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1354 = torch.ops.aten._weight_int4pack_mm.default(buf1353, arg748_1, 128, arg749_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg748_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg749_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1355 = buf1354
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1354
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_94], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1357 = torch.ops.aten._weight_int4pack_mm.default(buf1356, arg751_1, 128, arg752_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg751_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg752_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1358 = buf1357
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1357
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_94], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1360 = torch.ops.aten._weight_int4pack_mm.default(buf1359, arg754_1, 128, arg755_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg754_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg755_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1361 = buf1360
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1360
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1362 = reinterpret_tensor(buf1355, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1355  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_448], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1362, arg750_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg750_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1363 = reinterpret_tensor(buf1358, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1358  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_448], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1363, arg753_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg753_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1364 = reinterpret_tensor(buf1361, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1361  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_448], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1364, arg756_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg756_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_448], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1366 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1362, buf1363, buf1364, reinterpret_tensor(buf1365, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1362
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1363
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1364
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1367 = buf1366[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1366
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1371 = buf1353; del buf1353  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_451], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1367, buf1371, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_451], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1372 = torch.ops.aten._weight_int4pack_mm.default(buf1371, arg757_1, 128, arg758_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg757_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg758_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1373 = buf1372
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1372
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1377 = reinterpret_tensor(buf1367, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1367  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_190, hidden_states_453, hidden_states_454, mul_98, norm_hidden_states_94, norm_hidden_states_95], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1373, arg759_1, buf1352, arg735_1, buf9, arg14_1, buf1377, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1378 = buf1371; del buf1371  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_455], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1377, buf1378, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_455], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1379 = torch.ops.aten._weight_int4pack_mm.default(buf1378, arg760_1, 128, arg761_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg760_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg761_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1380 = buf1379
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1379
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1381 = buf1324; del buf1324  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_458], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf1380, arg762_1, buf1381, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg762_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1380
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_458], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1382 = torch.ops.aten._weight_int4pack_mm.default(buf1381, arg763_1, 128, arg764_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg763_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg764_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1383 = buf1382
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1382
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1384 = reinterpret_tensor(buf1383, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1383  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1388 = buf1377; del buf1377  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_194, ff_output_23, hidden_states_453, hidden_states_454, hidden_states_459, mul_100, norm_hidden_states_96, norm_hidden_states_97], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1384, arg735_1, buf9, arg14_1, arg765_1, buf1373, arg759_1, buf1352, arg766_1, buf1388, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg735_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg759_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg765_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1352
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1373
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1389 = buf1378; del buf1378  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1392 = buf1335; del buf1335  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1395 = buf1332; del buf1332  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_96, query_96, value_96], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf1388, buf1389, buf1392, buf1395, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1388
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_96], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1390 = torch.ops.aten._weight_int4pack_mm.default(buf1389, arg767_1, 128, arg768_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg767_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg768_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1391 = buf1390
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1390
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_96], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1393 = torch.ops.aten._weight_int4pack_mm.default(buf1392, arg770_1, 128, arg771_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg770_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg771_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1394 = buf1393
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1393
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_96], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1396 = torch.ops.aten._weight_int4pack_mm.default(buf1395, arg773_1, 128, arg774_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg773_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg774_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1397 = buf1396
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1396
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1398 = reinterpret_tensor(buf1391, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1391  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_460], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1398, arg769_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg769_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1399 = reinterpret_tensor(buf1394, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1394  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_460], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1399, arg772_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg772_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1400 = reinterpret_tensor(buf1397, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1397  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_460], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1400, arg775_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg775_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_460], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1401 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1398, buf1399, buf1400, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1398
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1399
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1400
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1402 = buf1401[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1401
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1406 = buf1395; del buf1395  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_463], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf1402, buf1406, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1402
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_463], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1407 = torch.ops.aten._weight_int4pack_mm.default(buf1406, arg776_1, 128, arg777_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg776_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg777_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1408 = buf1407
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1407
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1409 = reinterpret_tensor(buf1408, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1408  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_24, hidden_states_465, hidden_states_466], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf1409, arg766_1, buf9, arg14_1, arg778_1, buf1384, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg778_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1384
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1410 = buf1406; del buf1406  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_98], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1409, buf1410, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_98], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1411 = torch.ops.aten._weight_int4pack_mm.default(buf1410, arg779_1, 128, arg780_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg779_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg780_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1412 = buf1411
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1411
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1413 = buf1359; del buf1359  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1416 = buf1356; del buf1356  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1470 = buf1302; del buf1302  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1473 = buf1299; del buf1299  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_102, key_98, value_102, value_98], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf1413, buf1416, buf1470, buf1473, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_98], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1414 = torch.ops.aten._weight_int4pack_mm.default(buf1413, arg782_1, 128, arg783_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg782_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg783_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1415 = buf1414
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1414
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_98], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1417 = torch.ops.aten._weight_int4pack_mm.default(buf1416, arg785_1, 128, arg786_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg785_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg786_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1418 = buf1417
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1417
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1419 = reinterpret_tensor(buf1412, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1412  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_467], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1419, arg781_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg781_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1420 = reinterpret_tensor(buf1415, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1415  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_467], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1420, arg784_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg784_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1421 = reinterpret_tensor(buf1418, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1418  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_467], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1421, arg787_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg787_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1422 = buf1365; del buf1365  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1479 = buf1308; del buf1308  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_467, hidden_states_486], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf1422, buf1479, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_467], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1423 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1419, buf1420, buf1421, reinterpret_tensor(buf1422, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1419
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1420
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1421
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1424 = buf1423[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1423
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1428 = buf1410; del buf1410  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_470], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1424, buf1428, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_470], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1429 = torch.ops.aten._weight_int4pack_mm.default(buf1428, arg788_1, 128, arg789_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg788_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg789_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1430 = buf1429
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1429
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1434 = reinterpret_tensor(buf1424, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1424  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_198, hidden_states_472, hidden_states_473, mul_102, norm_hidden_states_98, norm_hidden_states_99], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1430, arg790_1, buf1409, arg766_1, buf9, arg14_1, buf1434, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1435 = buf1428; del buf1428  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_474], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1434, buf1435, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_474], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1436 = torch.ops.aten._weight_int4pack_mm.default(buf1435, arg791_1, 128, arg792_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg791_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg792_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1437 = buf1436
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1436
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1438 = buf1381; del buf1381  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_477], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf1437, arg793_1, buf1438, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg793_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1437
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_477], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1439 = torch.ops.aten._weight_int4pack_mm.default(buf1438, arg794_1, 128, arg795_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg794_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg795_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1440 = buf1439
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1439
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1441 = reinterpret_tensor(buf1440, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1440  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1445 = buf1434; del buf1434  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_202, ff_output_24, hidden_states_472, hidden_states_473, hidden_states_478, mul_104, norm_hidden_states_100, norm_hidden_states_101], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1441, arg766_1, buf9, arg14_1, arg796_1, buf1430, arg790_1, buf1409, arg797_1, buf1445, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg766_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg790_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg796_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1409
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1430
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1446 = buf1435; del buf1435  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1449 = buf1392; del buf1392  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1452 = buf1389; del buf1389  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_100, query_100, value_100], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf1445, buf1446, buf1449, buf1452, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1445
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_100], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1447 = torch.ops.aten._weight_int4pack_mm.default(buf1446, arg798_1, 128, arg799_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg798_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg799_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1448 = buf1447
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1447
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_100], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1450 = torch.ops.aten._weight_int4pack_mm.default(buf1449, arg801_1, 128, arg802_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg801_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg802_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1451 = buf1450
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1450
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_100], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1453 = torch.ops.aten._weight_int4pack_mm.default(buf1452, arg804_1, 128, arg805_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg804_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg805_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1454 = buf1453
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1453
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1455 = reinterpret_tensor(buf1448, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1448  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_479], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1455, arg800_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg800_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1456 = reinterpret_tensor(buf1451, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1451  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_479], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1456, arg803_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg803_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1457 = reinterpret_tensor(buf1454, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1454  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_479], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1457, arg806_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg806_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_479], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1458 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1455, buf1456, buf1457, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1455
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1456
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1457
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1459 = buf1458[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1458
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1463 = buf1452; del buf1452  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_482], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf1459, buf1463, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1459
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_482], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1464 = torch.ops.aten._weight_int4pack_mm.default(buf1463, arg807_1, 128, arg808_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg807_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg808_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1465 = buf1464
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1464
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1466 = reinterpret_tensor(buf1465, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1465  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_25, hidden_states_484, hidden_states_485], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf1466, arg797_1, buf9, arg14_1, arg809_1, buf1441, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg809_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1441
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1467 = buf1463; del buf1463  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_102], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1466, buf1467, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_102], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1468 = torch.ops.aten._weight_int4pack_mm.default(buf1467, arg810_1, 128, arg811_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg810_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg811_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1469 = buf1468
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1468
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_102], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1471 = torch.ops.aten._weight_int4pack_mm.default(buf1470, arg813_1, 128, arg814_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg813_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg814_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1472 = buf1471
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1471
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_102], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1474 = torch.ops.aten._weight_int4pack_mm.default(buf1473, arg816_1, 128, arg817_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg816_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg817_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1475 = buf1474
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1474
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1476 = reinterpret_tensor(buf1469, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1469  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_486], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1476, arg812_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg812_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1477 = reinterpret_tensor(buf1472, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1472  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_486], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1477, arg815_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg815_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1478 = reinterpret_tensor(buf1475, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1475  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_486], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1478, arg818_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg818_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_486], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1480 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1476, buf1477, buf1478, reinterpret_tensor(buf1479, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1476
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1477
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1478
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1481 = buf1480[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1480
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1485 = buf1467; del buf1467  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_489], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1481, buf1485, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_489], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1486 = torch.ops.aten._weight_int4pack_mm.default(buf1485, arg819_1, 128, arg820_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg819_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg820_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1487 = buf1486
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1486
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1491 = reinterpret_tensor(buf1481, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1481  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_206, hidden_states_491, hidden_states_492, mul_106, norm_hidden_states_102, norm_hidden_states_103], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1487, arg821_1, buf1466, arg797_1, buf9, arg14_1, buf1491, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1492 = buf1485; del buf1485  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_493], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1491, buf1492, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_493], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1493 = torch.ops.aten._weight_int4pack_mm.default(buf1492, arg822_1, 128, arg823_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg822_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg823_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1494 = buf1493
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1493
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1495 = buf1438; del buf1438  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_496], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf1494, arg824_1, buf1495, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg824_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1494
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_496], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1496 = torch.ops.aten._weight_int4pack_mm.default(buf1495, arg825_1, 128, arg826_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg825_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg826_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1497 = buf1496
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1496
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1498 = reinterpret_tensor(buf1497, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1497  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1502 = buf1491; del buf1491  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_210, ff_output_25, hidden_states_491, hidden_states_492, hidden_states_497, mul_108, norm_hidden_states_104, norm_hidden_states_105], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1498, arg797_1, buf9, arg14_1, arg827_1, buf1487, arg821_1, buf1466, arg828_1, buf1502, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg797_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg821_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg827_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1466
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1487
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1503 = buf1492; del buf1492  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1506 = buf1449; del buf1449  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1509 = buf1446; del buf1446  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_104, query_104, value_104], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf1502, buf1503, buf1506, buf1509, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1502
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_104], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1504 = torch.ops.aten._weight_int4pack_mm.default(buf1503, arg829_1, 128, arg830_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg829_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg830_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1505 = buf1504
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1504
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_104], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1507 = torch.ops.aten._weight_int4pack_mm.default(buf1506, arg832_1, 128, arg833_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg832_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg833_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1508 = buf1507
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1507
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_104], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1510 = torch.ops.aten._weight_int4pack_mm.default(buf1509, arg835_1, 128, arg836_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg835_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg836_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1511 = buf1510
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1510
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1512 = reinterpret_tensor(buf1505, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1505  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_498], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1512, arg831_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg831_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1513 = reinterpret_tensor(buf1508, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1508  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_498], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1513, arg834_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg834_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1514 = reinterpret_tensor(buf1511, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1511  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_498], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1514, arg837_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg837_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_498], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1515 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1512, buf1513, buf1514, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1512
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1513
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1514
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1516 = buf1515[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1515
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1520 = buf1509; del buf1509  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_501], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf1516, buf1520, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1516
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_501], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1521 = torch.ops.aten._weight_int4pack_mm.default(buf1520, arg838_1, 128, arg839_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg838_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg839_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1522 = buf1521
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1521
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1523 = reinterpret_tensor(buf1522, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1522  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_26, hidden_states_503, hidden_states_504], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf1523, arg828_1, buf9, arg14_1, arg840_1, buf1498, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg840_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1498
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1524 = buf1520; del buf1520  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_106], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1523, buf1524, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_106], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1525 = torch.ops.aten._weight_int4pack_mm.default(buf1524, arg841_1, 128, arg842_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg841_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg842_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1526 = buf1525
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1525
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1527 = buf1473; del buf1473  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1530 = buf1470; del buf1470  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1584 = buf1416; del buf1416  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1587 = buf1413; del buf1413  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_106, key_110, value_106, value_110], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf1527, buf1530, buf1584, buf1587, 1228800, grid=grid(1228800), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg21_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf44
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_106], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1528 = torch.ops.aten._weight_int4pack_mm.default(buf1527, arg844_1, 128, arg845_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg844_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg845_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1527
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1529 = buf1528
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1528
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_106], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1531 = torch.ops.aten._weight_int4pack_mm.default(buf1530, arg847_1, 128, arg848_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg847_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg848_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1530
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1532 = buf1531
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1531
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1533 = reinterpret_tensor(buf1526, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1526  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_505], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1533, arg843_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg843_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1534 = reinterpret_tensor(buf1529, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1529  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_505], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1534, arg846_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg846_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1535 = reinterpret_tensor(buf1532, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1532  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_505], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1535, arg849_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg849_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1536 = buf1479; del buf1479  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1593 = buf1422; del buf1422  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_505, hidden_states_524], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf1536, buf1593, 9728, grid=grid(9728), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg0_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_505], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1537 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1533, buf1534, buf1535, reinterpret_tensor(buf1536, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1533
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1534
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1535
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1536
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1538 = buf1537[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1537
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1542 = buf1524; del buf1524  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_508], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1538, buf1542, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_508], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1543 = torch.ops.aten._weight_int4pack_mm.default(buf1542, arg850_1, 128, arg851_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg850_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg851_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1544 = buf1543
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1543
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1548 = reinterpret_tensor(buf1538, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1538  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_214, hidden_states_510, hidden_states_511, mul_110, norm_hidden_states_106, norm_hidden_states_107], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1544, arg852_1, buf1523, arg828_1, buf9, arg14_1, buf1548, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1549 = buf1542; del buf1542  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_512], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1548, buf1549, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_512], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1550 = torch.ops.aten._weight_int4pack_mm.default(buf1549, arg853_1, 128, arg854_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg853_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg854_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1551 = buf1550
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1550
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1552 = buf1495; del buf1495  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_515], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf1551, arg855_1, buf1552, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg855_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1551
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_515], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1553 = torch.ops.aten._weight_int4pack_mm.default(buf1552, arg856_1, 128, arg857_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg856_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg857_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1554 = buf1553
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1553
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1555 = reinterpret_tensor(buf1554, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1554  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1559 = buf1548; del buf1548  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_218, ff_output_26, hidden_states_510, hidden_states_511, hidden_states_516, mul_112, norm_hidden_states_108, norm_hidden_states_109], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1555, arg828_1, buf9, arg14_1, arg858_1, buf1544, arg852_1, buf1523, arg859_1, buf1559, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg828_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg852_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg858_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1523
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1544
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1560 = buf1549; del buf1549  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1563 = buf1506; del buf1506  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1566 = buf1503; del buf1503  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_108, query_108, value_108], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_6.run(buf1559, buf1560, buf1563, buf1566, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1559
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_108], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1561 = torch.ops.aten._weight_int4pack_mm.default(buf1560, arg860_1, 128, arg861_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg860_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg861_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1560
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1562 = buf1561
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1561
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_108], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1564 = torch.ops.aten._weight_int4pack_mm.default(buf1563, arg863_1, 128, arg864_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg863_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg864_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1563
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1565 = buf1564
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1564
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_108], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1567 = torch.ops.aten._weight_int4pack_mm.default(buf1566, arg866_1, 128, arg867_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg866_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg867_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1568 = buf1567
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1567
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1569 = reinterpret_tensor(buf1562, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1562  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_517], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1569, arg862_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg862_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1570 = reinterpret_tensor(buf1565, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1565  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_517], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1570, arg865_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg865_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1571 = reinterpret_tensor(buf1568, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1568  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_517], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1571, arg868_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg868_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_517], Original ATen: [aten._scaled_dot_product_cudnn_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1572 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1569, buf1570, buf1571, None, False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1569
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1570
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1571
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1573 = buf1572[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1572
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1577 = buf1566; del buf1566  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_520], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_8.run(buf1573, buf1577, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1573
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_520], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1578 = torch.ops.aten._weight_int4pack_mm.default(buf1577, arg869_1, 128, arg870_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg869_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg870_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1579 = buf1578
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1578
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1580 = reinterpret_tensor(buf1579, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1579  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [attn_output_27, hidden_states_522, hidden_states_523], Original ATen: [aten.add, aten.div, aten.mul]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_add_div_mul_18.run(buf1580, arg859_1, buf9, arg14_1, arg871_1, buf1555, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg871_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1555
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1581 = buf1577; del buf1577  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_110], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1580, buf1581, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [query_110], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1582 = torch.ops.aten._weight_int4pack_mm.default(buf1581, arg872_1, 128, arg873_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg872_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg873_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1583 = buf1582
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1582
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [key_110], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1585 = torch.ops.aten._weight_int4pack_mm.default(buf1584, arg875_1, 128, arg876_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg875_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg876_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1584
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1586 = buf1585
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1585
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [value_110], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1588 = torch.ops.aten._weight_int4pack_mm.default(buf1587, arg878_1, 128, arg879_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg878_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg879_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1587
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1589 = buf1588
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1588
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1590 = reinterpret_tensor(buf1583, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1583  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_524], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1590, arg874_1, 9437184, grid=grid(9437184), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg874_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1591 = reinterpret_tensor(buf1586, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1586  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_524], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1591, arg877_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg877_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1592 = reinterpret_tensor(buf1589, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1589  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_524], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1592, arg880_1, 691200, grid=grid(691200), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg880_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_524], Original ATen: [aten._scaled_dot_product_efficient_attention]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1594 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1590, buf1591, buf1592, reinterpret_tensor(buf1593, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1590
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1591
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1592
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1593
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1595 = buf1594[0]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1594
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1599 = buf1581; del buf1581  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_527], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1595, buf1599, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_527], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1600 = torch.ops.aten._weight_int4pack_mm.default(buf1599, arg881_1, 128, arg882_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg881_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg882_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1601 = buf1600
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1600
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1605 = reinterpret_tensor(buf1595, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1595  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [add_222, hidden_states_529, hidden_states_530, mul_114, norm_hidden_states_110, norm_hidden_states_111], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1601, arg883_1, buf1580, arg859_1, buf9, arg14_1, buf1605, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1606 = buf1599; del buf1599  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_531], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_10.run(buf1605, buf1606, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1605
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_531], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1607 = torch.ops.aten._weight_int4pack_mm.default(buf1606, arg884_1, 128, arg885_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg884_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg885_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1608 = buf1607
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1607
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1609 = buf1552; del buf1552  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_534], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_16.run(buf1608, arg886_1, buf1609, 41943040, grid=grid(41943040), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg886_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1608
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_534], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1610 = torch.ops.aten._weight_int4pack_mm.default(buf1609, arg887_1, 128, arg888_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg887_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg888_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1609
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1611 = buf1610
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1610
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1612 = reinterpret_tensor(buf1611, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1611  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1613 = empty_strided_cuda((2, 4096, 1), (4096, 1, 8192), torch.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1614 = empty_strided_cuda((2, 4096, 1), (4096, 1, 8192), torch.float32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [ff_output_27, hidden_states_529, hidden_states_530, hidden_states_535, hidden_states_536], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_red_fused_add_div_mul_native_layer_norm_19.run(buf1612, arg859_1, buf9, arg14_1, arg889_1, buf1601, arg883_1, buf1580, buf1613, buf1614, 8192, 1152, grid=grid(8192), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg14_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg859_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg883_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg889_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1580
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1601
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf9
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1616 = buf1606; del buf1606  # reuse
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_538], Original ATen: [aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_constant_pad_nd_20.run(buf1612, buf1613, buf1614, arg890_1, buf6, arg11_1, buf1616, 16777216, grid=grid(16777216), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg11_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg890_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1612
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1613
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1614
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf6
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [hidden_states_538], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1617 = torch.ops.aten._weight_int4pack_mm.default(buf1616, arg891_1, 128, arg892_1)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg891_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg892_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1616
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1618 = buf1617
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1617
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         buf1619 = empty_strided_cuda((2, 8, 64, 2, 64, 2), (131072, 16384, 256, 128, 2, 1), torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         # Source Nodes: [output], Original ATen: [aten.clone]
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         triton_poi_fused_clone_21.run(buf1618, arg893_1, buf1619, 16, 16384, grid=grid(16, 16384), stream=stream0)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del arg893_1
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]         del buf1618
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     return (reinterpret_tensor(buf1619, (2, 8, 128, 128), (131072, 16384, 128, 1), 0), )
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def benchmark_compiled_module(times=10, repeat=10):
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     from torch._dynamo.testing import rand_strided
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     from torch._inductor.utils import print_performance
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg0_1 = rand_strided((2, 300), (300, 1), device='cuda:0', dtype=torch.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg1_1 = rand_strided((2, 4, 128, 128), (65536, 16384, 128, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg2_1 = rand_strided((1152, 4, 2, 2), (16, 1, 8, 4), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg3_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg4_1 = rand_strided((1, 4096, 1152), (4718592, 1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg5_1 = rand_strided((2, ), (0, ), device='cuda:0', dtype=torch.int64)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg6_1 = rand_strided((144, 8, 32, 4), (1024, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg7_1 = rand_strided((8, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg8_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg9_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg10_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg11_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg12_1 = rand_strided((864, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg13_1 = rand_strided((16, 6912, 2), (13824, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg14_1 = rand_strided((6912, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg15_1 = rand_strided((144, 32, 32, 4), (4096, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg16_1 = rand_strided((32, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg17_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg18_1 = rand_strided((2, 300, 4096), (1228800, 4096, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg19_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg20_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg21_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg22_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg23_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg24_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg25_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg26_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg27_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg28_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg29_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg30_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg31_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg32_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg33_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg34_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg35_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg36_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg37_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg38_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg39_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg40_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg41_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg42_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg43_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg44_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg45_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg46_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg47_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg48_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg49_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg50_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg51_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg52_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg53_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg54_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg55_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg56_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg57_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg58_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg59_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg60_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg61_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg62_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg63_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg64_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg65_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg66_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg67_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg68_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg69_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg70_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg71_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg72_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg73_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg74_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg75_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg76_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg77_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg78_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg79_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg80_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg81_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg82_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg83_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg84_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg85_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg86_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg87_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg88_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg89_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg90_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg91_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg92_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg93_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg94_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg95_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg96_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg97_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg98_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg99_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg100_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg101_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg102_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg103_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg104_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg105_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg106_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg107_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg108_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg109_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg110_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg111_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg112_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg113_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg114_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg115_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg116_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg117_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg118_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg119_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg120_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg121_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg122_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg123_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg124_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg125_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg126_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg127_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg128_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg129_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg130_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg131_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg132_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg133_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg134_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg135_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg136_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg137_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg138_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg139_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg140_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg141_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg142_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg143_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg144_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg145_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg146_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg147_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg148_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg149_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg150_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg151_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg152_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg153_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg154_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg155_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg156_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg157_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg158_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg159_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg160_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg161_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg162_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg163_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg164_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg165_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg166_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg167_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg168_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg169_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg170_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg171_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg172_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg173_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg174_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg175_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg176_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg177_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg178_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg179_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg180_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg181_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg182_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg183_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg184_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg185_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg186_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg187_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg188_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg189_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg190_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg191_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg192_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg193_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg194_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg195_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg196_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg197_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg198_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg199_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg200_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg201_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg202_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg203_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg204_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg205_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg206_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg207_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg208_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg209_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg210_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg211_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg212_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg213_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg214_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg215_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg216_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg217_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg218_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg219_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg220_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg221_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg222_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg223_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg224_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg225_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg226_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg227_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg228_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg229_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg230_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg231_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg232_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg233_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg234_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg235_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg236_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg237_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg238_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg239_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg240_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg241_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg242_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg243_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg244_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg245_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg246_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg247_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg248_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg249_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg250_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg251_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg252_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg253_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg254_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg255_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg256_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg257_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg258_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg259_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg260_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg261_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg262_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg263_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg264_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg265_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg266_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg267_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg268_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg269_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg270_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg271_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg272_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg273_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg274_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg275_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg276_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg277_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg278_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg279_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg280_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg281_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg282_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg283_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg284_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg285_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg286_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg287_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg288_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg289_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg290_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg291_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg292_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg293_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg294_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg295_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg296_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg297_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg298_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg299_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg300_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg301_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg302_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg303_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg304_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg305_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg306_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg307_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg308_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg309_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg310_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg311_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg312_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg313_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg314_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg315_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg316_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg317_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg318_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg319_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg320_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg321_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg322_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg323_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg324_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg325_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg326_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg327_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg328_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg329_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg330_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg331_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg332_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg333_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg334_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg335_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg336_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg337_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg338_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg339_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg340_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg341_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg342_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg343_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg344_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg345_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg346_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg347_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg348_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg349_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg350_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg351_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg352_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg353_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg354_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg355_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg356_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg357_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg358_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg359_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg360_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg361_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg362_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg363_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg364_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg365_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg366_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg367_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg368_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg369_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg370_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg371_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg372_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg373_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg374_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg375_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg376_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg377_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg378_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg379_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg380_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg381_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg382_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg383_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg384_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg385_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg386_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg387_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg388_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg389_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg390_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg391_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg392_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg393_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg394_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg395_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg396_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg397_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg398_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg399_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg400_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg401_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg402_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg403_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg404_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg405_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg406_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg407_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg408_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg409_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg410_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg411_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg412_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg413_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg414_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg415_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg416_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg417_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg418_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg419_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg420_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg421_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg422_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg423_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg424_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg425_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg426_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg427_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg428_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg429_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg430_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg431_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg432_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg433_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg434_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg435_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg436_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg437_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg438_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg439_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg440_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg441_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg442_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg443_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg444_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg445_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg446_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg447_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg448_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg449_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg450_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg451_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg452_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg453_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg454_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg455_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg456_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg457_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg458_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg459_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg460_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg461_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg462_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg463_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg464_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg465_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg466_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg467_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg468_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg469_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg470_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg471_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg472_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg473_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg474_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg475_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg476_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg477_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg478_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg479_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg480_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg481_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg482_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg483_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg484_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg485_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg486_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg487_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg488_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg489_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg490_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg491_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg492_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg493_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg494_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg495_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg496_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg497_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg498_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg499_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg500_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg501_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg502_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg503_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg504_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg505_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg506_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg507_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg508_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg509_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg510_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg511_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg512_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg513_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg514_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg515_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg516_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg517_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg518_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg519_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg520_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg521_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg522_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg523_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg524_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg525_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg526_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg527_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg528_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg529_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg530_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg531_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg532_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg533_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg534_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg535_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg536_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg537_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg538_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg539_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg540_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg541_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg542_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg543_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg544_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg545_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg546_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg547_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg548_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg549_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg550_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg551_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg552_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg553_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg554_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg555_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg556_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg557_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg558_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg559_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg560_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg561_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg562_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg563_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg564_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg565_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg566_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg567_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg568_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg569_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg570_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg571_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg572_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg573_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg574_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg575_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg576_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg577_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg578_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg579_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg580_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg581_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg582_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg583_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg584_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg585_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg586_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg587_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg588_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg589_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg590_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg591_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg592_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg593_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg594_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg595_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg596_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg597_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg598_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg599_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg600_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg601_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg602_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg603_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg604_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg605_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg606_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg607_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg608_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg609_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg610_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg611_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg612_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg613_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg614_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg615_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg616_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg617_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg618_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg619_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg620_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg621_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg622_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg623_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg624_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg625_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg626_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg627_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg628_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg629_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg630_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg631_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg632_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg633_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg634_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg635_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg636_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg637_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg638_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg639_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg640_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg641_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg642_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg643_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg644_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg645_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg646_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg647_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg648_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg649_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg650_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg651_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg652_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg653_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg654_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg655_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg656_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg657_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg658_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg659_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg660_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg661_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg662_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg663_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg664_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg665_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg666_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg667_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg668_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg669_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg670_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg671_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg672_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg673_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg674_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg675_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg676_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg677_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg678_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg679_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg680_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg681_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg682_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg683_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg684_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg685_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg686_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg687_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg688_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg689_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg690_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg691_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg692_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg693_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg694_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg695_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg696_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg697_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg698_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg699_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg700_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg701_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg702_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg703_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg704_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg705_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg706_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg707_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg708_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg709_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg710_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg711_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg712_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg713_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg714_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg715_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg716_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg717_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg718_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg719_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg720_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg721_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg722_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg723_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg724_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg725_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg726_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg727_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg728_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg729_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg730_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg731_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg732_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg733_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg734_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg735_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg736_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg737_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg738_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg739_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg740_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg741_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg742_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg743_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg744_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg745_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg746_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg747_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg748_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg749_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg750_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg751_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg752_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg753_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg754_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg755_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg756_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg757_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg758_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg759_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg760_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg761_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg762_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg763_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg764_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg765_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg766_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg767_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg768_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg769_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg770_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg771_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg772_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg773_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg774_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg775_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg776_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg777_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg778_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg779_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg780_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg781_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg782_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg783_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg784_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg785_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg786_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg787_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg788_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg789_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg790_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg791_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg792_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg793_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg794_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg795_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg796_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg797_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg798_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg799_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg800_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg801_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg802_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg803_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg804_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg805_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg806_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg807_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg808_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg809_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg810_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg811_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg812_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg813_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg814_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg815_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg816_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg817_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg818_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg819_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg820_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg821_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg822_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg823_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg824_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg825_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg826_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg827_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg828_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg829_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg830_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg831_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg832_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg833_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg834_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg835_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg836_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg837_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg838_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg839_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg840_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg841_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg842_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg843_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg844_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg845_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg846_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg847_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg848_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg849_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg850_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg851_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg852_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg853_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg854_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg855_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg856_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg857_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg858_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg859_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg860_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg861_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg862_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg863_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg864_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg865_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg866_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg867_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg868_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg869_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg870_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg871_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg872_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg873_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg874_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg875_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg876_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg877_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg878_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg879_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg880_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg881_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg882_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg883_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg884_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg885_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg886_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg887_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg888_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg889_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg890_1 = rand_strided((2, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg891_1 = rand_strided((4, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg892_1 = rand_strided((16, 32, 2), (64, 2, 1), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     arg893_1 = rand_strided((32, ), (1, ), device='cuda:0', dtype=torch.bfloat16)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     fn = lambda: call([arg0_1, arg1_1, arg2_1, arg3_1, arg4_1, arg5_1, arg6_1, arg7_1, arg8_1, arg9_1, arg10_1, arg11_1, arg12_1, arg13_1, arg14_1, arg15_1, arg16_1, arg17_1, arg18_1, arg19_1, arg20_1, arg21_1, arg22_1, arg23_1, arg24_1, arg25_1, arg26_1, arg27_1, arg28_1, arg29_1, arg30_1, arg31_1, arg32_1, arg33_1, arg34_1, arg35_1, arg36_1, arg37_1, arg38_1, arg39_1, arg40_1, arg41_1, arg42_1, arg43_1, arg44_1, arg45_1, arg46_1, arg47_1, arg48_1, arg49_1, arg50_1, arg51_1, arg52_1, arg53_1, arg54_1, arg55_1, arg56_1, arg57_1, arg58_1, arg59_1, arg60_1, arg61_1, arg62_1, arg63_1, arg64_1, arg65_1, arg66_1, arg67_1, arg68_1, arg69_1, arg70_1, arg71_1, arg72_1, arg73_1, arg74_1, arg75_1, arg76_1, arg77_1, arg78_1, arg79_1, arg80_1, arg81_1, arg82_1, arg83_1, arg84_1, arg85_1, arg86_1, arg87_1, arg88_1, arg89_1, arg90_1, arg91_1, arg92_1, arg93_1, arg94_1, arg95_1, arg96_1, arg97_1, arg98_1, arg99_1, arg100_1, arg101_1, arg102_1, arg103_1, arg104_1, arg105_1, arg106_1, arg107_1, arg108_1, arg109_1, arg110_1, arg111_1, arg112_1, arg113_1, arg114_1, arg115_1, arg116_1, arg117_1, arg118_1, arg119_1, arg120_1, arg121_1, arg122_1, arg123_1, arg124_1, arg125_1, arg126_1, arg127_1, arg128_1, arg129_1, arg130_1, arg131_1, arg132_1, arg133_1, arg134_1, arg135_1, arg136_1, arg137_1, arg138_1, arg139_1, arg140_1, arg141_1, arg142_1, arg143_1, arg144_1, arg145_1, arg146_1, arg147_1, arg148_1, arg149_1, arg150_1, arg151_1, arg152_1, arg153_1, arg154_1, arg155_1, arg156_1, arg157_1, arg158_1, arg159_1, arg160_1, arg161_1, arg162_1, arg163_1, arg164_1, arg165_1, arg166_1, arg167_1, arg168_1, arg169_1, arg170_1, arg171_1, arg172_1, arg173_1, arg174_1, arg175_1, arg176_1, arg177_1, arg178_1, arg179_1, arg180_1, arg181_1, arg182_1, arg183_1, arg184_1, arg185_1, arg186_1, arg187_1, arg188_1, arg189_1, arg190_1, arg191_1, arg192_1, arg193_1, arg194_1, arg195_1, arg196_1, arg197_1, arg198_1, arg199_1, arg200_1, arg201_1, arg202_1, arg203_1, arg204_1, arg205_1, arg206_1, arg207_1, arg208_1, arg209_1, arg210_1, arg211_1, arg212_1, arg213_1, arg214_1, arg215_1, arg216_1, arg217_1, arg218_1, arg219_1, arg220_1, arg221_1, arg222_1, arg223_1, arg224_1, arg225_1, arg226_1, arg227_1, arg228_1, arg229_1, arg230_1, arg231_1, arg232_1, arg233_1, arg234_1, arg235_1, arg236_1, arg237_1, arg238_1, arg239_1, arg240_1, arg241_1, arg242_1, arg243_1, arg244_1, arg245_1, arg246_1, arg247_1, arg248_1, arg249_1, arg250_1, arg251_1, arg252_1, arg253_1, arg254_1, arg255_1, arg256_1, arg257_1, arg258_1, arg259_1, arg260_1, arg261_1, arg262_1, arg263_1, arg264_1, arg265_1, arg266_1, arg267_1, arg268_1, arg269_1, arg270_1, arg271_1, arg272_1, arg273_1, arg274_1, arg275_1, arg276_1, arg277_1, arg278_1, arg279_1, arg280_1, arg281_1, arg282_1, arg283_1, arg284_1, arg285_1, arg286_1, arg287_1, arg288_1, arg289_1, arg290_1, arg291_1, arg292_1, arg293_1, arg294_1, arg295_1, arg296_1, arg297_1, arg298_1, arg299_1, arg300_1, arg301_1, arg302_1, arg303_1, arg304_1, arg305_1, arg306_1, arg307_1, arg308_1, arg309_1, arg310_1, arg311_1, arg312_1, arg313_1, arg314_1, arg315_1, arg316_1, arg317_1, arg318_1, arg319_1, arg320_1, arg321_1, arg322_1, arg323_1, arg324_1, arg325_1, arg326_1, arg327_1, arg328_1, arg329_1, arg330_1, arg331_1, arg332_1, arg333_1, arg334_1, arg335_1, arg336_1, arg337_1, arg338_1, arg339_1, arg340_1, arg341_1, arg342_1, arg343_1, arg344_1, arg345_1, arg346_1, arg347_1, arg348_1, arg349_1, arg350_1, arg351_1, arg352_1, arg353_1, arg354_1, arg355_1, arg356_1, arg357_1, arg358_1, arg359_1, arg360_1, arg361_1, arg362_1, arg363_1, arg364_1, arg365_1, arg366_1, arg367_1, arg368_1, arg369_1, arg370_1, arg371_1, arg372_1, arg373_1, arg374_1, arg375_1, arg376_1, arg377_1, arg378_1, arg379_1, arg380_1, arg381_1, arg382_1, arg383_1, arg384_1, arg385_1, arg386_1, arg387_1, arg388_1, arg389_1, arg390_1, arg391_1, arg392_1, arg393_1, arg394_1, arg395_1, arg396_1, arg397_1, arg398_1, arg399_1, arg400_1, arg401_1, arg402_1, arg403_1, arg404_1, arg405_1, arg406_1, arg407_1, arg408_1, arg409_1, arg410_1, arg411_1, arg412_1, arg413_1, arg414_1, arg415_1, arg416_1, arg417_1, arg418_1, arg419_1, arg420_1, arg421_1, arg422_1, arg423_1, arg424_1, arg425_1, arg426_1, arg427_1, arg428_1, arg429_1, arg430_1, arg431_1, arg432_1, arg433_1, arg434_1, arg435_1, arg436_1, arg437_1, arg438_1, arg439_1, arg440_1, arg441_1, arg442_1, arg443_1, arg444_1, arg445_1, arg446_1, arg447_1, arg448_1, arg449_1, arg450_1, arg451_1, arg452_1, arg453_1, arg454_1, arg455_1, arg456_1, arg457_1, arg458_1, arg459_1, arg460_1, arg461_1, arg462_1, arg463_1, arg464_1, arg465_1, arg466_1, arg467_1, arg468_1, arg469_1, arg470_1, arg471_1, arg472_1, arg473_1, arg474_1, arg475_1, arg476_1, arg477_1, arg478_1, arg479_1, arg480_1, arg481_1, arg482_1, arg483_1, arg484_1, arg485_1, arg486_1, arg487_1, arg488_1, arg489_1, arg490_1, arg491_1, arg492_1, arg493_1, arg494_1, arg495_1, arg496_1, arg497_1, arg498_1, arg499_1, arg500_1, arg501_1, arg502_1, arg503_1, arg504_1, arg505_1, arg506_1, arg507_1, arg508_1, arg509_1, arg510_1, arg511_1, arg512_1, arg513_1, arg514_1, arg515_1, arg516_1, arg517_1, arg518_1, arg519_1, arg520_1, arg521_1, arg522_1, arg523_1, arg524_1, arg525_1, arg526_1, arg527_1, arg528_1, arg529_1, arg530_1, arg531_1, arg532_1, arg533_1, arg534_1, arg535_1, arg536_1, arg537_1, arg538_1, arg539_1, arg540_1, arg541_1, arg542_1, arg543_1, arg544_1, arg545_1, arg546_1, arg547_1, arg548_1, arg549_1, arg550_1, arg551_1, arg552_1, arg553_1, arg554_1, arg555_1, arg556_1, arg557_1, arg558_1, arg559_1, arg560_1, arg561_1, arg562_1, arg563_1, arg564_1, arg565_1, arg566_1, arg567_1, arg568_1, arg569_1, arg570_1, arg571_1, arg572_1, arg573_1, arg574_1, arg575_1, arg576_1, arg577_1, arg578_1, arg579_1, arg580_1, arg581_1, arg582_1, arg583_1, arg584_1, arg585_1, arg586_1, arg587_1, arg588_1, arg589_1, arg590_1, arg591_1, arg592_1, arg593_1, arg594_1, arg595_1, arg596_1, arg597_1, arg598_1, arg599_1, arg600_1, arg601_1, arg602_1, arg603_1, arg604_1, arg605_1, arg606_1, arg607_1, arg608_1, arg609_1, arg610_1, arg611_1, arg612_1, arg613_1, arg614_1, arg615_1, arg616_1, arg617_1, arg618_1, arg619_1, arg620_1, arg621_1, arg622_1, arg623_1, arg624_1, arg625_1, arg626_1, arg627_1, arg628_1, arg629_1, arg630_1, arg631_1, arg632_1, arg633_1, arg634_1, arg635_1, arg636_1, arg637_1, arg638_1, arg639_1, arg640_1, arg641_1, arg642_1, arg643_1, arg644_1, arg645_1, arg646_1, arg647_1, arg648_1, arg649_1, arg650_1, arg651_1, arg652_1, arg653_1, arg654_1, arg655_1, arg656_1, arg657_1, arg658_1, arg659_1, arg660_1, arg661_1, arg662_1, arg663_1, arg664_1, arg665_1, arg666_1, arg667_1, arg668_1, arg669_1, arg670_1, arg671_1, arg672_1, arg673_1, arg674_1, arg675_1, arg676_1, arg677_1, arg678_1, arg679_1, arg680_1, arg681_1, arg682_1, arg683_1, arg684_1, arg685_1, arg686_1, arg687_1, arg688_1, arg689_1, arg690_1, arg691_1, arg692_1, arg693_1, arg694_1, arg695_1, arg696_1, arg697_1, arg698_1, arg699_1, arg700_1, arg701_1, arg702_1, arg703_1, arg704_1, arg705_1, arg706_1, arg707_1, arg708_1, arg709_1, arg710_1, arg711_1, arg712_1, arg713_1, arg714_1, arg715_1, arg716_1, arg717_1, arg718_1, arg719_1, arg720_1, arg721_1, arg722_1, arg723_1, arg724_1, arg725_1, arg726_1, arg727_1, arg728_1, arg729_1, arg730_1, arg731_1, arg732_1, arg733_1, arg734_1, arg735_1, arg736_1, arg737_1, arg738_1, arg739_1, arg740_1, arg741_1, arg742_1, arg743_1, arg744_1, arg745_1, arg746_1, arg747_1, arg748_1, arg749_1, arg750_1, arg751_1, arg752_1, arg753_1, arg754_1, arg755_1, arg756_1, arg757_1, arg758_1, arg759_1, arg760_1, arg761_1, arg762_1, arg763_1, arg764_1, arg765_1, arg766_1, arg767_1, arg768_1, arg769_1, arg770_1, arg771_1, arg772_1, arg773_1, arg774_1, arg775_1, arg776_1, arg777_1, arg778_1, arg779_1, arg780_1, arg781_1, arg782_1, arg783_1, arg784_1, arg785_1, arg786_1, arg787_1, arg788_1, arg789_1, arg790_1, arg791_1, arg792_1, arg793_1, arg794_1, arg795_1, arg796_1, arg797_1, arg798_1, arg799_1, arg800_1, arg801_1, arg802_1, arg803_1, arg804_1, arg805_1, arg806_1, arg807_1, arg808_1, arg809_1, arg810_1, arg811_1, arg812_1, arg813_1, arg814_1, arg815_1, arg816_1, arg817_1, arg818_1, arg819_1, arg820_1, arg821_1, arg822_1, arg823_1, arg824_1, arg825_1, arg826_1, arg827_1, arg828_1, arg829_1, arg830_1, arg831_1, arg832_1, arg833_1, arg834_1, arg835_1, arg836_1, arg837_1, arg838_1, arg839_1, arg840_1, arg841_1, arg842_1, arg843_1, arg844_1, arg845_1, arg846_1, arg847_1, arg848_1, arg849_1, arg850_1, arg851_1, arg852_1, arg853_1, arg854_1, arg855_1, arg856_1, arg857_1, arg858_1, arg859_1, arg860_1, arg861_1, arg862_1, arg863_1, arg864_1, arg865_1, arg866_1, arg867_1, arg868_1, arg869_1, arg870_1, arg871_1, arg872_1, arg873_1, arg874_1, arg875_1, arg876_1, arg877_1, arg878_1, arg879_1, arg880_1, arg881_1, arg882_1, arg883_1, arg884_1, arg885_1, arg886_1, arg887_1, arg888_1, arg889_1, arg890_1, arg891_1, arg892_1, arg893_1])
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     return print_performance(fn, times=times, repeat=repeat)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] if __name__ == "__main__":
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     from torch._inductor.wrapper_benchmark import compiled_module_main
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code]     compiled_module_main('None', benchmark_compiled_module)
V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] 
I0808 07:30:31.449268 1266882 torch/_inductor/graph.py:1814] [0/0] [__output_code] Output code written to: /tmp/torchinductor_sayak/zb/czbvhwljs3kusciezfqmyq4h44hfd5dma5sfliqylshsi5zrrsli.py