class GraphModule(torch.nn.Module): def forward(self, L_v_: "f16[1, 256, 68, 68][1183744, 4624, 68, 1]cuda:0", L_self_modules_relative_emb_k_parameters_weight_: "f32[225, 256, 1, 1][256, 1, 1, 1]cuda:0", L_self_modules_relative_emb_k_parameters_bias_: "f32[225][1]cuda:0", L_q_: "f16[1, 256, 68, 68][1183744, 4624, 68, 1]cuda:0", L_k_: "f16[1, 256, 68, 68][1183744, 4624, 68, 1]cuda:0"): l_v_ = L_v_ l_self_modules_relative_emb_k_parameters_weight_ = L_self_modules_relative_emb_k_parameters_weight_ l_self_modules_relative_emb_k_parameters_bias_ = L_self_modules_relative_emb_k_parameters_bias_ l_q_ = L_q_ l_k_ = L_k_ # File: /workspace/networks/layers/attention.py:316 in forward, code: n, c, h, w = v.size() size = l_v_.size() getitem_2: "Sym(68)" = size[2] getitem_3: "Sym(68)" = size[3]; size = None # File: /opt/conda/lib/python3.11/site-packages/torch/nn/modules/conv.py:453 in _conv_forward, code: return F.conv2d(input, weight, bias, self.stride, relative_emb: "f16[1, 225, 68, 68][1040400, 4624, 68, 1]cuda:0" = torch.conv2d(l_q_, l_self_modules_relative_emb_k_parameters_weight_, l_self_modules_relative_emb_k_parameters_bias_, (1, 1), (0, 0), (1, 1), 1); l_self_modules_relative_emb_k_parameters_weight_ = l_self_modules_relative_emb_k_parameters_bias_ = None # File: /workspace/networks/layers/attention.py:335 in forward, code: q = q / self.T q: "f16[1, 256, 68, 68][1183744, 4624, 68, 1]cuda:0" = l_q_ / 16.0; l_q_ = None # File: /workspace/networks/layers/attention.py:337 in forward, code: q = q.view(-1, self.d_att, h, w) q_1: "f16[1, 256, 68, 68][1183744, 4624, 68, 1]cuda:0" = q.view(-1, 256, getitem_2, getitem_3); q = None # File: /workspace/networks/layers/attention.py:338 in forward, code: k = k.view(-1, self.d_att, h, w) k: "f16[1, 256, 68, 68][1183744, 4624, 68, 1]cuda:0" = l_k_.view(-1, 256, getitem_2, getitem_3); l_k_ = None # File: /workspace/networks/layers/attention.py:339 in forward, code: v = v.view(-1, self.num_head, hidden_dim, h * w) mul: "Sym(4624)" = getitem_2 * getitem_3 v: "f16[1, 1, 256, 4624][1183744, 1183744, 4624, 1]cuda:0" = l_v_.view(-1, 1, 256, mul); l_v_ = mul = None # File: /workspace/networks/layers/attention.py:343 in forward, code: h * w) mul_1: "Sym(4624)" = getitem_2 * getitem_3 # File: /workspace/networks/layers/attention.py:341 in forward, code: relative_emb = relative_emb.view(n, self.num_head, relative_emb_1: "f16[1, 1, 225, 4624][1040400, 1040400, 4624, 1]cuda:0" = relative_emb.view(1, 1, 225, mul_1); relative_emb = mul_1 = None return (q_1, k, v, getitem_2, getitem_3, relative_emb_1)