diff --git a/compare.py b/compare.py index 3da58e7..91de438 100644 --- a/compare.py +++ b/compare.py @@ -1,6 +1,9 @@ -import numpy as np import json -datum = np.load('dump.npz.npy') + +import numpy as np + +#datum = np.load('dump.npz.npy') +datum = np.load('dump_so400m.npy') with open('dump.json','r') as rr: @@ -29,25 +32,15 @@ def cosine_sim(emb_in_1, emb_in_2): arr_in_deepstream = np.asarray([y for _,y in emb_dict.items()]) - normed = np.divide(datum.T, np.linalg.norm(datum, axis=1)).T print('_________________________') + +print(len(emb_dict)) +print(len(datum)) for fr, emb in emb_dict.items(): emb1 = np.linalg.norm(emb) emb2 = np.linalg.norm(datum[fr]) -# print( cosine_sim(emb, datum[fr])) + print( cosine_sim(emb, datum[fr])) -print('Deepstream and Actual norm') -print(np.max(np.dot(arr_in_deepstream, normed.T),axis=1)) - -print('_________________________') -for dat in datum: -# print(cosine_sim(dat, datum[0])) - pass - - -#print(cosine_sim(datum[fr], datum[fr+1])) - -#print(cosine_sim(emb_dict[fr], emb_dict[fr+1])) diff --git a/deepstream_obj_det_pre_queue.py b/deepstream_obj_det_pre_queue.py index 4fa53ea..ae72082 100755 --- a/deepstream_obj_det_pre_queue.py +++ b/deepstream_obj_det_pre_queue.py @@ -178,7 +178,7 @@ def embed_results_probe(pad, info, u_data, list_add, frame_num=0): if True: for i in range(tensor_meta.num_output_layers): layer = pyds.get_nvds_LayerInfo(tensor_meta, i) - if layer.layerName == "embedding": + if layer.layerName == "output": ptr = ctypes.cast( pyds.get_ptr(layer.buffer), ctypes.POINTER(ctypes.c_float) @@ -395,8 +395,8 @@ if True: streammux_embed.set_property("width", target_width_embed) streammux_embed.set_property("height", target_height_embed) streammux_embed.set_property("batched-push-timeout", MUXER_BATCH_TIMEOUT_USEC) - streammux_embed.set_property("enable-padding", 1) - streammux_embed.set_property("batch-size", 4) + streammux_embed.set_property("enable-padding", 0) + streammux_embed.set_property("batch-size", 1) nugget_embed = Gst.ElementFactory.make("nvinfer", "primary-inference") nugget_embed.set_property( @@ -426,7 +426,7 @@ if True: # capsfilter.link(tee) nvvidconv.link(tee) - if True: + if False: pipeline.add(queue_detect) pipeline.add(streammux_detect) pipeline.add(nugget_detector) @@ -442,7 +442,7 @@ if True: os.environ["GST_DEBUG_DUMP_DOT_DIR"] = "/tmp" os.putenv("GST_DEBUG_DUMP_DIR_DIR", "/tmp") - if False: + if True: pipeline.add(queue_embed) pipeline.add(streammux_embed) @@ -522,7 +522,13 @@ if True: pass # cleanup pipeline.set_state(Gst.State.NULL) -# return detector_list, embed_list + + # return detector_list, embed_list\\ + out = [detector_list, embed_list ] + import json + with open("dump.json", "w") as ff: + json.dump([out[0], out[1]], ff) + sys.exit() if __name__ == "__main__": @@ -537,6 +543,3 @@ if __name__ == "__main__": import json - with open("dump.json", "w") as ff: - json.dump([out[0], out[1]], ff) - sys.exit() diff --git a/dump.json b/dump.json index e6f9476..bfa346c 100644 --- a/dump.json +++ b/dump.json @@ -1 +1 @@ -[[], []] \ No newline at end of file +[[], [{"frame_number": 0, "vector": [-0.144287109375, 0.34619140625, 0.00439453125, -0.173828125, -0.5966796875, 0.309814453125, 0.01568603515625, -0.481201171875, -0.0948486328125, 0.0217132568359375, -0.122314453125, 0.006591796875, 0.27587890625, 0.19482421875, 0.0050048828125, 0.2958984375, 0.14892578125, -0.18505859375, -0.06292724609375, -0.4794921875, -0.509765625, -6.3359375, -0.3466796875, -0.66845703125, -0.0172271728515625, -0.1317138671875, -0.2060546875, -0.80810546875, 0.4619140625, -0.188720703125, -0.9482421875, -0.67236328125, -0.78955078125, 0.04510498046875, -0.07110595703125, 0.100830078125, 0.2890625, -0.6376953125, -0.06585693359375, -0.466552734375, -0.044921875, 0.264404296875, 0.189453125, 0.014892578125, 0.93798828125, -0.324951171875, -0.39306640625, -0.216796875, -0.552734375, -0.15478515625, 0.080078125, -0.255859375, 0.071044921875, 0.479248046875, 0.2197265625, 0.016510009765625, 0.37060546875, -0.259765625, 0.27587890625, -0.0067138671875, -0.058837890625, 0.2156982421875, 0.307861328125, 0.182861328125, -0.214599609375, -0.27880859375, -0.50537109375, -0.0867919921875, 0.167236328125, 0.138427734375, 0.93359375, -0.09716796875, 0.025390625, -0.037841796875, -0.45166015625, 0.8544921875, -0.322021484375, -0.43359375, 0.166015625, -0.68505859375, 1.1796875, 0.309814453125, 0.60107421875, 0.5458984375, -0.273193359375, 0.022216796875, 0.2093505859375, 0.0594482421875, 0.1884765625, -0.55859375, -0.1566162109375, -0.133544921875, 0.086669921875, -0.166015625, 0.278564453125, 0.312744140625, 0.1090087890625, 0.6025390625, 0.0284881591796875, -0.6748046875, 0.736328125, 0.275634765625, -0.42236328125, -0.61328125, -0.0272979736328125, -0.150634765625, 0.08831787109375, -0.185302734375, -0.18359375, -0.062255859375, -0.08819580078125, 0.16259765625, 0.0574951171875, -0.4404296875, 0.1881103515625, -0.5595703125, -0.200439453125, -0.143310546875, 0.31591796875, 0.5087890625, 0.0389404296875, -0.262939453125, -0.0745849609375, -0.7099609375, -0.083984375, 0.2091064453125, -0.177978515625, 0.356689453125, 0.20068359375, 0.302001953125, 0.014404296875, -0.40771484375, 0.0810546875, 0.07098388671875, 0.52001953125, -0.046875, 0.0501708984375, 0.74951171875, 0.6103515625, 0.068603515625, 0.07763671875, 0.314453125, 0.1527099609375, -0.0750732421875, -0.265869140625, 0.0950927734375, 0.00946044921875, -0.380615234375, 0.2020263671875, 0.994140625, -0.1480712890625, -1.5, 0.0721435546875, -0.42529296875, 0.463134765625, -0.205078125, -0.291015625, 0.9501953125, 0.314697265625, -0.420654296875, -0.218994140625, 0.204833984375, 0.468017578125, 0.31689453125, -0.3681640625, -0.283935546875, -0.060302734375, 0.71484375, 0.077880859375, 0.0662841796875, 0.164306640625, 0.456298828125, -0.189697265625, 0.67578125, -0.2138671875, 1.0439453125, -1.0693359375, 0.075927734375, -0.0113525390625, -0.08203125, 0.033447265625, 0.224853515625, -0.1781005859375, -0.0117034912109375, 0.053466796875, 0.00872802734375, -0.038330078125, 0.21630859375, -0.1292724609375, 0.17431640625, 0.15625, 0.18017578125, 0.4404296875, -0.3623046875, 0.478515625, -0.2176513671875, 0.450439453125, -0.26806640625, -0.31982421875, 1.14453125, 0.476806640625, -0.04931640625, -0.25830078125, 0.0660400390625, -0.06549072265625, 0.07818603515625, 0.60791015625, 0.060455322265625, -0.6416015625, -0.2032470703125, 0.1781005859375, -0.08447265625, 0.279052734375, -0.818359375, -0.54345703125, -0.509765625, 0.438720703125, 1.0205078125, 0.04656982421875, -0.45263671875, -0.40478515625, -0.56103515625, -0.1619873046875, -0.58203125, -0.1572265625, 0.16064453125, -0.0999755859375, -0.29345703125, 0.7548828125, 0.02215576171875, -0.08770751953125, 0.29248046875, 0.046630859375, 0.263916015625, -0.325927734375, -0.58837890625, 0.771484375, -0.1285400390625, -0.83642578125, -0.085693359375, 0.74755859375, 0.264404296875, 0.020263671875, 0.134033203125, -0.1842041015625, 1.0703125, -0.206787109375, 0.0289306640625, -0.454833984375, -0.06591796875, 0.0185546875, -0.04962158203125, 0.292724609375, 0.0458984375, -0.724609375, -0.15185546875, -0.157958984375, 0.0081787109375, 0.4951171875, 0.47021484375, -0.0322265625, -0.2099609375, -0.101318359375, 0.0135498046875, -0.26806640625, 0.037353515625, -0.8115234375, -0.2197265625, 0.5390625, -0.517578125, 0.067138671875, 0.34375, 0.34619140625, 0.210205078125, -0.173095703125, -0.65283203125, -0.05999755859375, 0.1270751953125, -0.3291015625, -0.40869140625, -0.37646484375, 0.6689453125, -0.425048828125, 0.173095703125, 0.365234375, 0.368896484375, -0.1585693359375, 0.4921875, -0.58935546875, 1.142578125, -0.001617431640625, 0.042266845703125, 0.0562744140625, -0.401123046875, 0.08966064453125, 0.19384765625, 0.155517578125, 0.3837890625, 0.6494140625, -0.258056640625, -0.0258331298828125, 0.7421875, -0.0255889892578125, -0.071533203125, -0.3271484375, 0.340087890625, 0.37890625, -0.681640625, 0.1402587890625, -0.06298828125, 0.333984375, -0.38671875, -0.021240234375, -0.233642578125, -0.314453125, 0.374755859375, -0.02874755859375, 0.058349609375, -0.25, 0.178955078125, -0.921875, 0.143310546875, 0.094482421875, -0.4814453125, -0.8779296875, 0.6708984375, -0.42919921875, 0.304443359375, -0.58837890625, -0.040374755859375, -0.5810546875, -0.05224609375, -0.2100830078125, -0.4541015625, 0.205810546875, 0.2276611328125, 0.0712890625, -0.5947265625, 0.03131103515625, -0.0003662109375, -0.6142578125, 0.104248046875, -0.25, -0.7158203125, -0.47900390625, 0.2392578125, -0.52490234375, 0.41259765625, 0.226806640625, 0.0302734375, -0.36376953125, -0.322265625, 0.2342529296875, 0.136474609375, 0.422607421875, 0.324462890625, 0.2449951171875, -0.1802978515625, -0.26806640625, -0.2098388671875, -0.173828125, 0.0830078125, 0.48974609375, 0.2734375, -0.5986328125, -0.4970703125, -0.03369140625, 0.2056884765625, 0.0657958984375, -0.0830078125, -0.07122802734375, 0.2724609375, -0.16748046875, 0.329833984375, 0.30078125, -0.0274658203125, 0.39697265625, -0.2724609375, 0.22265625, -0.50390625, 0.0010986328125, -0.366455078125, 0.0960693359375, 0.7119140625, 0.2230224609375, -0.018310546875, 0.00433349609375, -0.134521484375, -0.255126953125, 0.52783203125, -0.63671875, -0.255126953125, 0.24462890625, 0.370849609375, 1.0751953125, 0.6943359375, -0.082275390625, -0.630859375, 0.012451171875, -0.43798828125, 0.265869140625, 0.1561279296875, -0.12158203125, -0.43505859375, -0.1932373046875, 0.050048828125, -0.355224609375, 0.06787109375, 0.330078125, -0.0401611328125, 0.210205078125, -0.82861328125, -0.72900390625, -0.35009765625, -0.404296875, 0.51416015625, 0.68359375, -0.270751953125, 1.037109375, -0.252685546875, 0.10546875, -0.043701171875, -0.285400390625, -0.1146240234375, -0.0382080078125, -0.5595703125, -1.3935546875, -0.1370849609375, 0.0955810546875, 0.17431640625, 0.0574951171875, -6.3359375, -0.4072265625, -0.072265625, -0.7451171875, -0.1761474609375, -0.1002197265625, -1.201171875, -0.66552734375, -0.2978515625, -0.020751953125, 0.625, 0.7001953125, -0.239013671875, -0.033447265625, 0.1900634765625, 0.265380859375, -0.0133056640625, -0.0849609375, 0.187255859375, 0.00341796875, 0.17822265625, -0.1458740234375, 0.0240478515625, 0.37060546875, -0.319091796875, -0.1563720703125, 0.414306640625, -0.6240234375, -0.6064453125, -0.062255859375, 0.180908203125, -0.283203125, -0.3349609375, -0.47802734375, -0.0648193359375, 0.52880859375, 0.763671875, -0.34619140625, -0.37353515625, 0.252197265625, -0.96435546875, 0.3095703125, -0.0323486328125, -0.0311279296875, 0.248779296875, 0.30712890625, -0.0234375, -0.73291015625, -0.77978515625, 0.1361083984375, 0.04931640625, -1.091796875, -0.0263671875, 0.39306640625, -0.03271484375, 0.24072265625, -0.09173583984375, 0.136962890625, 0.12744140625, 0.220947265625, 0.5693359375, 0.37060546875, 0.4091796875, -0.486083984375, 0.04736328125, 0.0830078125, 0.0362548828125, 0.076904296875, 0.093017578125, -0.0531005859375, 0.04034423828125, 0.6005859375, -0.1148681640625, -0.216796875, -0.09765625, -0.031005859375, 0.1549072265625, -0.18310546875, 0.102294921875, 0.233642578125, 0.0579833984375, 0.2017822265625, 0.4013671875, -0.17626953125, -0.1527099609375, -0.09716796875, -0.321044921875, -0.5703125, -0.1619873046875, 0.77587890625, 0.23779296875, -0.18310546875, 0.03607177734375, 0.21923828125, 1.3486328125, 0.15478515625, 0.16357421875, 0.1649169921875, 0.1307373046875, -0.338623046875, -0.5576171875, 0.05322265625, -0.07568359375, 0.3828125, 0.32080078125, 0.00927734375, -0.03955078125, 0.38330078125, -0.0704345703125, 0.7275390625, 0.14306640625, 0.332275390625, 0.451171875, -0.233154296875, 0.459228515625, -1.548828125, -0.30126953125, -0.1995849609375, 0.1387939453125, 0.0050048828125, -0.333740234375, 0.771484375, 0.227783203125, -0.110595703125, 0.099853515625, 0.344970703125, -0.08251953125, -0.23779296875, -0.9150390625, -0.490966796875, -0.264892578125, -0.412353515625, -0.958984375, 0.28271484375, 0.321044921875, 0.56201171875, -0.29931640625, -0.0872802734375, 0.08380126953125, -0.0189208984375, 0.38623046875, -0.205078125, 0.119140625, 0.19580078125, 0.287841796875, -0.328125, 0.21044921875, -0.31640625, -0.467041015625, -0.28076171875, 0.06231689453125, -0.1435546875, -0.0926513671875, -0.056396484375, -0.095947265625, 0.037109375, 0.454345703125, -0.32373046875, 0.2205810546875, 0.424560546875, 0.02850341796875, -0.2178955078125, 0.06207275390625, 0.1080322265625, 0.083740234375, -0.048095703125, -0.1937255859375, -1.4765625, -0.4853515625, 0.193115234375, -0.070556640625, -0.348876953125, -0.31298828125, -0.16796875, -0.225830078125, -0.079345703125, -0.267333984375, -0.2349853515625, -0.9091796875, 0.28076171875, 0.39111328125, -0.1884765625, -0.104736328125, -0.0970458984375, -0.53076171875, 0.257568359375, 0.239990234375, -0.0853271484375, -0.2861328125, 0.010498046875, -0.060302734375, 0.0714111328125, 0.053955078125, -0.220947265625, -0.0533447265625, 0.039947509765625, 0.31591796875, 0.326416015625, -0.0458984375, 0.190673828125, 0.0479736328125, -0.353271484375, 0.73779296875, 0.02227783203125, -0.086669921875, -0.318115234375, 1.3212890625, 0.1328125, -0.262939453125, 0.0438232421875, -0.49560546875, 0.35888671875, -0.216064453125, -0.223876953125, 0.320068359375, 0.193603515625, 0.025634765625, 0.218017578125, 0.08038330078125, 0.2247314453125, -0.367431640625, 0.3232421875, 0.234375, -0.669921875, 0.19140625, 0.29150390625, -0.1085205078125, -0.2744140625, -0.2427978515625, -0.32275390625, -0.8388671875, 0.0081787109375, -0.13916015625, -0.07537841796875, -0.2012939453125, 0.106201171875, -0.073974609375, 0.43798828125, -0.402099609375, -1.4052734375, -0.146240234375, -0.1116943359375, 0.209716796875, 0.2666015625, -0.05474853515625, 0.380859375, 0.197998046875, -0.180419921875, 0.2303466796875, -0.03656005859375, 0.1656494140625, 0.417236328125, 0.01904296875, 0.126220703125, -0.2113037109375, 0.29638671875, 0.21240234375, -0.4130859375, -0.595703125, -0.207763671875, -0.386962890625, 0.146484375, 0.26806640625, -0.208740234375, 0.763671875, -0.0841064453125, 0.07373046875, -0.8330078125, 0.07373046875, 0.0191650390625, 0.18603515625, -0.6259765625, 0.349365234375, -0.55029296875, 0.375, -0.04949951171875, 0.194091796875, -0.10699462890625, -0.537109375, -0.62109375, 0.08990478515625, -0.43505859375, 0.40234375, -0.363525390625, 0.9052734375, -0.31689453125, 0.103271484375, -0.08984375, -0.0364990234375, -0.416015625, 0.1044921875, 1.333984375, -0.007904052734375, 0.2247314453125, -0.0699462890625, 0.3359375, -1.0, -0.2626953125, 0.513671875, 0.02685546875, -0.25146484375, 0.55859375, 0.154296875, -0.014404296875, 0.21044921875, 0.0496826171875, 0.209228515625, -0.0882568359375, 0.43359375, -0.5478515625, 0.017333984375, -0.6064453125, 0.0828857421875, 0.1187744140625, -0.423828125, -0.01654052734375, -0.0391845703125, -0.0911865234375, 0.092041015625, -0.927734375, -0.256591796875, 0.323974609375, -0.1719970703125, 0.2305908203125, 0.061767578125, -0.0762939453125, -0.8837890625, 0.05224609375, 0.32373046875, 0.01812744140625, 0.0208740234375, -0.29833984375, -0.5283203125, 0.6005859375, -0.39697265625, 0.00469970703125, -0.62255859375, -0.02685546875, -0.3662109375, 0.13525390625, 0.064453125, 0.0706787109375, 0.0496826171875, 0.1688232421875, -2.296875, 0.34423828125, -0.155517578125, 0.245849609375, 0.60986328125, -0.0250244140625, -0.339111328125, -0.22021484375, 0.12646484375, -0.13720703125, 0.18359375, -0.00830078125, 0.083984375, 0.529296875, 0.42529296875, 0.296630859375, 0.002685546875, 0.256591796875, 0.634765625, -0.2666015625, 0.36279296875, 0.10540771484375, 0.1060791015625, -0.3056640625, 0.19775390625, 0.228759765625, 0.25390625, -0.7470703125, 0.51953125, 0.0797119140625, -0.142822265625, 0.025146484375, -0.533203125, 0.00537109375, -0.0369873046875, 0.3203125, -0.287353515625, -0.008026123046875, -0.216064453125, -0.2255859375, 0.402587890625, -0.052490234375, -0.242919921875, -0.646484375, -0.436279296875, 0.334228515625, 0.061279296875, -0.1717529296875, 0.330078125, -0.1331787109375, 0.333251953125, -0.39208984375, -0.33056640625, 0.152587890625, -0.2203369140625, -0.4453125, -0.646484375, -0.68017578125, -0.02459716796875, 0.8408203125, 0.53076171875, -0.314208984375, -0.0169677734375, 0.65673828125, -0.64453125, 0.23583984375, -0.62939453125, -0.716796875, -0.52587890625, 0.11767578125, 0.093017578125, -0.311767578125, -0.314697265625, 0.1251220703125, -0.2353515625, 0.1591796875, 0.490966796875, 0.154541015625, 0.1802978515625, -0.035400390625, -0.52978515625, 0.076171875, -0.079345703125, -0.1019287109375, 0.74853515625, 0.1494140625, -0.328125, 0.5078125, -0.6416015625, 0.48046875, -0.41796875, 0.5634765625, -0.12103271484375, 0.056396484375, 0.1243896484375, -0.98193359375, -0.125732421875, 0.07843017578125, 0.2135009765625, -0.24609375, -0.51025390625, 0.0306396484375, 1.09375, 0.4248046875, -0.7197265625, -0.268798828125, 0.78662109375, -0.190673828125, -0.210693359375, -0.169189453125, -0.37890625, 0.188720703125, -0.202392578125, -0.460693359375, 0.406982421875, 0.41650390625, -0.356201171875, -0.5068359375, 0.073974609375, 0.0626220703125, -0.171142578125, 0.1502685546875, -0.1795654296875, -0.1241455078125, 0.271484375, 0.2578125, 0.355712890625, 0.35205078125, 0.17626953125, -0.13720703125, 0.2275390625, -0.0340576171875, 0.35546875, 0.369873046875, 0.46533203125, -0.2236328125, -0.107421875, 0.11083984375, -0.275634765625, -0.01904296875, -0.17431640625, 0.364990234375, 0.35107421875, -0.32568359375, 0.427978515625, -2.09765625, 0.50146484375, 0.399658203125, 0.126220703125, -0.58740234375, -0.1378173828125, 0.07421875, -0.165771484375, 0.0665283203125, 0.194580078125, 0.359130859375, -0.0751953125, 0.7021484375, -0.344970703125, 0.09765625, 0.64208984375, 0.26123046875, 0.24462890625, -0.22802734375, -0.3857421875, 0.162841796875, 0.5146484375, 0.137451171875, -0.042572021484375, -0.275390625, 0.228515625, 0.1055908203125, 0.9814453125, 0.291015625, 0.30322265625, -0.269775390625, 0.328369140625, 0.151123046875, 0.04168701171875, 0.6689453125, -0.67431640625, -0.307861328125, 0.5966796875, -0.60546875, 0.051513671875, 0.7041015625, -0.46435546875, -0.25048828125, -0.177001953125, -0.3720703125, -0.19921875, 0.079345703125, -0.88134765625, -0.12109375, 0.18310546875, 0.0548095703125, 0.6328125, -0.25390625, 0.41455078125, -0.2734375, -0.2015380859375, -0.1231689453125, -0.09912109375, 0.05859375, -0.15234375, 0.142822265625, 0.08636474609375, 0.123046875, 0.386474609375, -0.13623046875, -0.172607421875, -0.40576171875, 0.00269317626953125, 0.49609375, 1.2119140625, -0.14013671875, -0.324462890625, -0.56201171875, 0.5546875, 0.38720703125, 0.0389404296875, 0.31689453125, -0.272705078125, -1.39453125, -0.199462890625, 0.0726318359375, -0.59716796875, 0.32421875, 0.662109375, -0.07586669921875, 0.17529296875, -0.10546875, 0.065673828125, 0.1181640625, -0.034637451171875, -0.49169921875, 0.29638671875, -0.138427734375, -0.469482421875, 0.240234375, -0.11962890625, -0.35205078125, -0.1959228515625, -0.22265625, -0.294189453125, -0.41748046875, 0.044677734375, 0.022216796875, -0.337158203125, -0.46533203125, -0.1337890625, -0.482421875, -0.11700439453125, 0.2330322265625, -0.03662109375, -0.343017578125, -0.20703125, -0.14599609375, -0.2308349609375, 0.102783203125, 0.48095703125, 0.1153564453125, 0.19189453125, -0.771484375, -0.32177734375, -0.5830078125, -0.1695556640625, 0.014892578125, 0.184814453125, 0.189208984375, 0.10400390625, 0.418701171875, -0.0128326416015625, 0.1751708984375, 0.01483154296875, -0.177734375, 0.42626953125, 0.2005615234375, 0.5830078125, -0.2109375, 0.041015625, -0.77294921875, -0.1273193359375, 0.0546875, 0.1248779296875, -0.2138671875, -0.07940673828125, -0.54296875, 0.3349609375, -1.4921875, 0.075439453125, -0.129638671875, 1.232421875, -0.1995849609375, 0.40869140625, 0.40185546875, 0.467529296875, -0.185546875, -0.0579833984375, 0.2734375, -0.050048828125, 0.423828125, 0.5859375, -0.55078125, -0.283203125, -0.090576171875, -0.1981201171875, -0.049713134765625, 0.113525390625, 0.444091796875, 0.0240478515625, -0.603515625, 0.169921875, -0.168701171875, -0.0537109375, -0.7900390625, 0.89990234375, 0.0279541015625, 0.0133056640625, -0.02197265625, -0.607421875, 0.425537109375, 0.2191162109375, 0.53125, -0.094970703125, -0.65087890625, 0.64501953125, 0.25732421875, -0.1239013671875, 1.10546875, 0.063232421875, -0.1181640625, -0.7392578125, -0.2529296875, -0.297119140625, -0.0299072265625, 0.171875, 0.048004150390625, 0.4384765625, -0.178466796875, 0.1790771484375, -0.0245361328125, -0.49169921875, -0.32568359375, 0.213134765625, -0.4052734375, -0.27197265625, 0.0838623046875, -0.41650390625, -0.135498046875, -0.09332275390625, -0.007080078125, 0.93701171875, 0.08538818359375, -0.309326171875, -0.0767822265625, 0.16748046875, -0.0413818359375, -0.0274658203125, 0.29931640625, -0.117431640625, 0.2861328125, 0.60693359375, -0.49853515625, 0.26611328125, -0.31982421875, -0.75390625, 0.1734619140625, 0.26708984375, 0.05029296875, 0.039794921875, 0.0654296875, 0.157470703125, 0.0457763671875, -0.1875, 0.40185546875, 0.671875]}, {"frame_number": 1, "vector": [-0.12237548828125, 0.332275390625, -0.000244140625, -0.18603515625, -0.6083984375, 0.303466796875, -0.010040283203125, -0.472412109375, -0.12451171875, 0.001739501953125, -0.150390625, 0.016845703125, 0.3154296875, 0.16455078125, -0.004638671875, 0.2734375, 0.135009765625, -0.158447265625, -0.05780029296875, -0.45947265625, -0.51416015625, -6.2421875, -0.32373046875, -0.68896484375, -0.006805419921875, -0.1334228515625, -0.205078125, -0.7802734375, 0.469970703125, -0.185302734375, -0.939453125, -0.65673828125, -0.80224609375, 0.0389404296875, -0.08197021484375, 0.1043701171875, 0.2841796875, -0.64990234375, -0.05975341796875, -0.46630859375, -0.0428466796875, 0.2484130859375, 0.21533203125, 0.02978515625, 0.92138671875, -0.33056640625, -0.411865234375, -0.2130126953125, -0.55224609375, -0.13525390625, 0.0699462890625, -0.2587890625, 0.0782470703125, 0.47900390625, 0.2197265625, 0.04022216796875, 0.3564453125, -0.26806640625, 0.2744140625, -0.006103515625, -0.07861328125, 0.21044921875, 0.312255859375, 0.185302734375, -0.2431640625, -0.3046875, -0.5009765625, -0.0821533203125, 0.149658203125, 0.125244140625, 0.935546875, -0.0911865234375, 0.0048828125, -0.020263671875, -0.455810546875, 0.8515625, -0.32373046875, -0.44140625, 0.166259765625, -0.72900390625, 1.171875, 0.31640625, 0.59228515625, 0.529296875, -0.259765625, 0.0338134765625, 0.2003173828125, 0.035400390625, 0.19189453125, -0.57958984375, -0.1710205078125, -0.11328125, 0.06982421875, -0.176025390625, 0.29736328125, 0.32861328125, 0.0814208984375, 0.59521484375, 0.03204345703125, -0.68115234375, 0.736328125, 0.2822265625, -0.41552734375, -0.61279296875, -0.033111572265625, -0.1331787109375, 0.060699462890625, -0.181640625, -0.161376953125, -0.06298828125, -0.072998046875, 0.16552734375, 0.0772705078125, -0.479736328125, 0.183349609375, -0.591796875, -0.207275390625, -0.1614990234375, 0.328369140625, 0.5126953125, 0.0428466796875, -0.25341796875, -0.0699462890625, -0.70068359375, -0.0880126953125, 0.203857421875, -0.1510009765625, 0.35986328125, 0.20654296875, 0.326416015625, -0.0040283203125, -0.39404296875, 0.065185546875, 0.06365966796875, 0.5654296875, -0.0360107421875, 0.0655517578125, 0.7587890625, 0.59130859375, 0.048095703125, 0.091552734375, 0.30126953125, 0.156982421875, -0.0723876953125, -0.254638671875, 0.0771484375, -0.004241943359375, -0.3642578125, 0.194091796875, 1.0107421875, -0.141845703125, -1.51953125, 0.066650390625, -0.42529296875, 0.444580078125, -0.198974609375, -0.29443359375, 0.9443359375, 0.30224609375, -0.42822265625, -0.223876953125, 0.240966796875, 0.45849609375, 0.324462890625, -0.36669921875, -0.29296875, -0.0411376953125, 0.708984375, 0.062255859375, 0.0899658203125, 0.1673583984375, 0.418701171875, -0.2200927734375, 0.65087890625, -0.229736328125, 1.0302734375, -1.095703125, 0.0771484375, -0.0008544921875, -0.08935546875, 0.0181884765625, 0.20751953125, -0.1929931640625, 0.005767822265625, 0.0606689453125, 0.04559326171875, -0.01806640625, 0.208740234375, -0.123291015625, 0.201904296875, 0.14404296875, 0.1748046875, 0.444091796875, -0.36181640625, 0.509765625, -0.2205810546875, 0.42431640625, -0.26318359375, -0.3369140625, 1.14453125, 0.47412109375, -0.03759765625, -0.2255859375, 0.082763671875, -0.059814453125, 0.0650634765625, 0.654296875, 0.02532958984375, -0.6474609375, -0.16552734375, 0.2047119140625, -0.060302734375, 0.271240234375, -0.8330078125, -0.5185546875, -0.52490234375, 0.481201171875, 1.013671875, 0.0193023681640625, -0.427001953125, -0.4169921875, -0.56201171875, -0.1666259765625, -0.57861328125, -0.169677734375, 0.18798828125, -0.10791015625, -0.27734375, 0.76171875, 0.02313232421875, -0.08453369140625, 0.299560546875, 0.056396484375, 0.2607421875, -0.33251953125, -0.576171875, 0.77392578125, -0.057373046875, -0.88427734375, -0.099609375, 0.7431640625, 0.241943359375, 0.045654296875, 0.1162109375, -0.17138671875, 1.046875, -0.1776123046875, 0.039306640625, -0.44970703125, -0.08154296875, -0.0009765625, -0.0546875, 0.30859375, 0.005859375, -0.7265625, -0.1419677734375, -0.159912109375, -0.01513671875, 0.5078125, 0.4658203125, -0.0216064453125, -0.177001953125, -0.101806640625, -0.0115966796875, -0.27783203125, 0.01983642578125, -0.8154296875, -0.2412109375, 0.5693359375, -0.509765625, 0.05322265625, 0.373779296875, 0.35595703125, 0.23681640625, -0.133544921875, -0.66455078125, -0.06085205078125, 0.12152099609375, -0.339111328125, -0.4091796875, -0.356689453125, 0.6640625, -0.3994140625, 0.155517578125, 0.36474609375, 0.369384765625, -0.159912109375, 0.497314453125, -0.5732421875, 1.130859375, -0.0023345947265625, 0.06878662109375, 0.0574951171875, -0.404296875, 0.0904541015625, 0.2030029296875, 0.139404296875, 0.396728515625, 0.63671875, -0.263427734375, -0.0165557861328125, 0.71630859375, -0.0225372314453125, -0.059814453125, -0.3369140625, 0.3564453125, 0.369140625, -0.681640625, 0.12359619140625, -0.063232421875, 0.359619140625, -0.369873046875, -0.02001953125, -0.278564453125, -0.3251953125, 0.3740234375, -0.04595947265625, 0.046630859375, -0.26318359375, 0.1943359375, -0.9111328125, 0.12353515625, 0.0953369140625, -0.466796875, -0.8994140625, 0.6669921875, -0.428466796875, 0.31640625, -0.6015625, -0.0491943359375, -0.57470703125, -0.0391845703125, -0.192626953125, -0.438232421875, 0.2158203125, 0.2303466796875, 0.06475830078125, -0.5888671875, 0.0345458984375, 0.0030517578125, -0.65185546875, 0.0777587890625, -0.272216796875, -0.7236328125, -0.484130859375, 0.230224609375, -0.48779296875, 0.3994140625, 0.208251953125, 0.0361328125, -0.353515625, -0.30224609375, 0.249267578125, 0.1513671875, 0.43603515625, 0.34423828125, 0.25146484375, -0.1776123046875, -0.27587890625, -0.200927734375, -0.1536865234375, 0.036865234375, 0.5009765625, 0.285888671875, -0.6279296875, -0.4765625, -0.041748046875, 0.2088623046875, 0.100341796875, -0.0782470703125, -0.06939697265625, 0.2666015625, -0.16455078125, 0.322998046875, 0.311279296875, -0.0501708984375, 0.396484375, -0.26171875, 0.19970703125, -0.5126953125, -0.015869140625, -0.361328125, 0.10107421875, 0.67578125, 0.2298583984375, -0.0255126953125, 0.00042724609375, -0.13671875, -0.2587890625, 0.5009765625, -0.6630859375, -0.231689453125, 0.281005859375, 0.334716796875, 1.05078125, 0.677734375, -0.0810546875, -0.626953125, -0.0252685546875, -0.451171875, 0.2548828125, 0.1611328125, -0.1243896484375, -0.43896484375, -0.18310546875, 0.025146484375, -0.360107421875, 0.07763671875, 0.34716796875, -0.0452880859375, 0.213623046875, -0.841796875, -0.7099609375, -0.33740234375, -0.39404296875, 0.5380859375, 0.685546875, -0.28369140625, 1.0625, -0.25244140625, 0.12139892578125, -0.04302978515625, -0.2421875, -0.1173095703125, -0.05084228515625, -0.5146484375, -1.388671875, -0.11273193359375, 0.110595703125, 0.1876220703125, 0.07147216796875, -6.25, -0.416015625, -0.04931640625, -0.72021484375, -0.1619873046875, -0.1156005859375, -1.197265625, -0.64208984375, -0.280517578125, 0.001220703125, 0.6337890625, 0.70068359375, -0.2301025390625, -0.03753662109375, 0.16943359375, 0.24267578125, 0.007080078125, -0.0755615234375, 0.1685791015625, 0.0234375, 0.1881103515625, -0.1357421875, 0.0599365234375, 0.373291015625, -0.317138671875, -0.1810302734375, 0.398193359375, -0.64453125, -0.61572265625, -0.08251953125, 0.181884765625, -0.29296875, -0.349609375, -0.443359375, -0.0772705078125, 0.5087890625, 0.76220703125, -0.322998046875, -0.41943359375, 0.2349853515625, -0.9853515625, 0.30517578125, -0.03125, -0.0299835205078125, 0.241943359375, 0.27392578125, -0.060882568359375, -0.712890625, -0.78466796875, 0.12744140625, 0.0694580078125, -1.0927734375, -0.041015625, 0.365234375, -0.02685546875, 0.2353515625, -0.09814453125, 0.1357421875, 0.14501953125, 0.228271484375, 0.5703125, 0.359375, 0.416748046875, -0.47412109375, 0.03369140625, 0.05810546875, -0.00457763671875, 0.094970703125, 0.0789794921875, -0.06396484375, 0.004913330078125, 0.63818359375, -0.155517578125, -0.2041015625, -0.107666015625, -0.045166015625, 0.15185546875, -0.1748046875, 0.075439453125, 0.2509765625, 0.05517578125, 0.206298828125, 0.380859375, -0.180419921875, -0.1612548828125, -0.101318359375, -0.332763671875, -0.56201171875, -0.144775390625, 0.7470703125, 0.22119140625, -0.171630859375, 0.059783935546875, 0.22607421875, 1.3232421875, 0.166748046875, 0.144287109375, 0.17333984375, 0.09912109375, -0.34423828125, -0.546875, 0.0361328125, -0.06787109375, 0.38623046875, 0.322265625, 0.042724609375, -0.04388427734375, 0.42333984375, -0.0897216796875, 0.7255859375, 0.1553955078125, 0.338623046875, 0.48095703125, -0.225830078125, 0.460693359375, -1.529296875, -0.332763671875, -0.2022705078125, 0.14599609375, -0.0107421875, -0.308349609375, 0.7763671875, 0.271484375, -0.0970458984375, 0.07568359375, 0.3369140625, -0.072509765625, -0.23046875, -0.90673828125, -0.443115234375, -0.26220703125, -0.410400390625, -0.9775390625, 0.27294921875, 0.3251953125, 0.53369140625, -0.302734375, -0.07366943359375, 0.0806884765625, -0.0394287109375, 0.373779296875, -0.2109375, 0.111083984375, 0.182861328125, 0.2978515625, -0.35302734375, 0.2174072265625, -0.29833984375, -0.4775390625, -0.29638671875, 0.0889892578125, -0.16748046875, -0.093994140625, -0.030029296875, -0.10797119140625, 0.0498046875, 0.454833984375, -0.335693359375, 0.21728515625, 0.3994140625, 0.0321044921875, -0.23095703125, 0.03289794921875, 0.1007080078125, 0.0767822265625, -0.056640625, -0.1942138671875, -1.478515625, -0.49072265625, 0.1849365234375, -0.05224609375, -0.3525390625, -0.342041015625, -0.193603515625, -0.2078857421875, -0.0859375, -0.278564453125, -0.208740234375, -0.90478515625, 0.286865234375, 0.3916015625, -0.183837890625, -0.144287109375, -0.1065673828125, -0.56591796875, 0.247802734375, 0.23486328125, -0.07672119140625, -0.2646484375, 0.00714111328125, -0.057373046875, 0.06341552734375, 0.06781005859375, -0.2255859375, -0.04119873046875, 0.0391845703125, 0.3623046875, 0.293212890625, -0.065673828125, 0.1734619140625, 0.0445556640625, -0.364501953125, 0.75, 0.0501708984375, -0.10498046875, -0.314697265625, 1.3447265625, 0.093994140625, -0.267333984375, 0.031005859375, -0.46728515625, 0.365966796875, -0.20703125, -0.21533203125, 0.351806640625, 0.178955078125, 0.00823974609375, 0.226318359375, 0.0751953125, 0.2188720703125, -0.35107421875, 0.333251953125, 0.228271484375, -0.66162109375, 0.17626953125, 0.287841796875, -0.104736328125, -0.283203125, -0.26953125, -0.29541015625, -0.8583984375, -0.002685546875, -0.10595703125, -0.046600341796875, -0.2369384765625, 0.10546875, -0.05859375, 0.4296875, -0.426513671875, -1.390625, -0.14453125, -0.12744140625, 0.222900390625, 0.2568359375, -0.0523681640625, 0.365478515625, 0.193603515625, -0.172607421875, 0.2491455078125, -0.03082275390625, 0.1688232421875, 0.42822265625, 0.026123046875, 0.1217041015625, -0.1898193359375, 0.30126953125, 0.18798828125, -0.401123046875, -0.59375, -0.208984375, -0.43798828125, 0.09033203125, 0.25146484375, -0.2392578125, 0.7978515625, -0.083984375, 0.099609375, -0.845703125, 0.03619384765625, -0.010498046875, 0.1983642578125, -0.62353515625, 0.328125, -0.5439453125, 0.375, -0.04388427734375, 0.197021484375, -0.11273193359375, -0.53564453125, -0.6328125, 0.109375, -0.423828125, 0.422607421875, -0.344482421875, 0.93115234375, -0.331787109375, 0.0689697265625, -0.10791015625, -0.0382080078125, -0.40673828125, 0.101806640625, 1.34375, 0.0087890625, 0.2388916015625, -0.0723876953125, 0.334228515625, -1.0390625, -0.25341796875, 0.52734375, 0.0177001953125, -0.244873046875, 0.5537109375, 0.11566162109375, 0.015625, 0.1923828125, 0.0711669921875, 0.213623046875, -0.0946044921875, 0.4326171875, -0.52490234375, 0.025146484375, -0.6044921875, 0.0440673828125, 0.1163330078125, -0.4384765625, -0.01776123046875, -0.05804443359375, -0.093505859375, 0.093505859375, -0.92431640625, -0.2421875, 0.28564453125, -0.16259765625, 0.22998046875, 0.0289306640625, -0.0699462890625, -0.859375, 0.0648193359375, 0.336669921875, -0.0009765625, 0.027099609375, -0.260498046875, -0.5400390625, 0.58251953125, -0.4130859375, 0.01739501953125, -0.61279296875, -0.0128173828125, -0.3798828125, 0.134765625, 0.0234375, 0.0728759765625, 0.062408447265625, 0.171630859375, -2.26171875, 0.337158203125, -0.1708984375, 0.2314453125, 0.6318359375, -0.0394287109375, -0.3349609375, -0.205078125, 0.17138671875, -0.121337890625, 0.18798828125, 0.021240234375, 0.068603515625, 0.5400390625, 0.444580078125, 0.3232421875, -0.0244140625, 0.242431640625, 0.64501953125, -0.27099609375, 0.364013671875, 0.10467529296875, 0.1123046875, -0.27734375, 0.199462890625, 0.2384033203125, 0.27490234375, -0.75390625, 0.5537109375, 0.0953369140625, -0.1795654296875, 0.0223388671875, -0.537109375, -0.00048828125, -0.02130126953125, 0.3330078125, -0.281005859375, -0.019622802734375, -0.21142578125, -0.225341796875, 0.42041015625, -0.049560546875, -0.26708984375, -0.625, -0.4345703125, 0.3212890625, 0.086181640625, -0.1668701171875, 0.328857421875, -0.1256103515625, 0.314453125, -0.38818359375, -0.328125, 0.1593017578125, -0.23193359375, -0.495849609375, -0.658203125, -0.64892578125, -0.03277587890625, 0.8173828125, 0.54931640625, -0.27392578125, -0.013671875, 0.66162109375, -0.65576171875, 0.256591796875, -0.630859375, -0.7314453125, -0.5, 0.0975341796875, 0.0955810546875, -0.333984375, -0.312255859375, 0.157470703125, -0.18896484375, 0.1513671875, 0.45263671875, 0.153076171875, 0.14697265625, -0.0360107421875, -0.509765625, 0.0771484375, -0.06494140625, -0.0858154296875, 0.77001953125, 0.128173828125, -0.3330078125, 0.4931640625, -0.67822265625, 0.47119140625, -0.391357421875, 0.5849609375, -0.1085205078125, 0.070556640625, 0.1307373046875, -0.9814453125, -0.131591796875, 0.056640625, 0.19482421875, -0.241455078125, -0.49560546875, 0.03729248046875, 1.08203125, 0.4111328125, -0.724609375, -0.253173828125, 0.76806640625, -0.192626953125, -0.190185546875, -0.149658203125, -0.365966796875, 0.193359375, -0.2061767578125, -0.46923828125, 0.42529296875, 0.42626953125, -0.36376953125, -0.52978515625, 0.083740234375, 0.0726318359375, -0.1544189453125, 0.14990234375, -0.187255859375, -0.1099853515625, 0.23388671875, 0.27734375, 0.380859375, 0.3623046875, 0.1884765625, -0.13720703125, 0.200439453125, -0.04742431640625, 0.3623046875, 0.367431640625, 0.47607421875, -0.25439453125, -0.078369140625, 0.120361328125, -0.296142578125, -0.00244140625, -0.1671142578125, 0.367431640625, 0.35107421875, -0.31005859375, 0.437744140625, -2.115234375, 0.5, 0.4033203125, 0.141845703125, -0.5634765625, -0.1190185546875, 0.0687255859375, -0.15869140625, 0.07708740234375, 0.20751953125, 0.3388671875, -0.079345703125, 0.73388671875, -0.359619140625, 0.091064453125, 0.6904296875, 0.258056640625, 0.2578125, -0.23388671875, -0.3837890625, 0.190185546875, 0.5244140625, 0.15966796875, -0.05462646484375, -0.25341796875, 0.227783203125, 0.107666015625, 0.97998046875, 0.292724609375, 0.323486328125, -0.248291015625, 0.326904296875, 0.14501953125, 0.0260009765625, 0.67822265625, -0.67578125, -0.30517578125, 0.59033203125, -0.607421875, 0.0509033203125, 0.7041015625, -0.47412109375, -0.2607421875, -0.157470703125, -0.39208984375, -0.236572265625, 0.1097412109375, -0.89208984375, -0.1351318359375, 0.180908203125, 0.061767578125, 0.6123046875, -0.25048828125, 0.39404296875, -0.28857421875, -0.21630859375, -0.150634765625, -0.074462890625, 0.080810546875, -0.143798828125, 0.1539306640625, 0.0594482421875, 0.1143798828125, 0.373046875, -0.1591796875, -0.173095703125, -0.37841796875, 0.00173187255859375, 0.49462890625, 1.2236328125, -0.1429443359375, -0.3046875, -0.5634765625, 0.5537109375, 0.37158203125, 0.0384521484375, 0.34326171875, -0.273681640625, -1.36328125, -0.164794921875, 0.070068359375, -0.59423828125, 0.3427734375, 0.66064453125, -0.06884765625, 0.195068359375, -0.1414794921875, 0.0599365234375, 0.1298828125, -0.02349853515625, -0.48828125, 0.2958984375, -0.14404296875, -0.473876953125, 0.219970703125, -0.12060546875, -0.33642578125, -0.1898193359375, -0.2242431640625, -0.276611328125, -0.40087890625, 0.041259765625, 0.0238037109375, -0.333984375, -0.49267578125, -0.13037109375, -0.48876953125, -0.1083984375, 0.214599609375, -0.02978515625, -0.3564453125, -0.236572265625, -0.164794921875, -0.21630859375, 0.1494140625, 0.5068359375, 0.1312255859375, 0.208251953125, -0.7421875, -0.325439453125, -0.5419921875, -0.1988525390625, 0.008056640625, 0.18408203125, 0.179443359375, 0.0953369140625, 0.42431640625, -0.023101806640625, 0.1751708984375, 0.01934814453125, -0.1773681640625, 0.42138671875, 0.2030029296875, 0.57568359375, -0.18212890625, 0.0517578125, -0.78466796875, -0.1322021484375, 0.0599365234375, 0.1356201171875, -0.2437744140625, -0.0955810546875, -0.51171875, 0.34521484375, -1.484375, 0.09075927734375, -0.137939453125, 1.224609375, -0.1910400390625, 0.39208984375, 0.38525390625, 0.455078125, -0.17822265625, -0.078857421875, 0.275390625, -0.031005859375, 0.42578125, 0.6064453125, -0.5732421875, -0.26904296875, -0.084716796875, -0.192138671875, -0.0560302734375, 0.123779296875, 0.453857421875, 0.02655029296875, -0.5810546875, 0.1593017578125, -0.167724609375, -0.076416015625, -0.767578125, 0.90869140625, 0.0252685546875, 0.022705078125, -0.0208740234375, -0.6337890625, 0.4375, 0.2127685546875, 0.5146484375, -0.074951171875, -0.6435546875, 0.68359375, 0.25830078125, -0.1190185546875, 1.103515625, 0.08978271484375, -0.0885009765625, -0.7578125, -0.253173828125, -0.2900390625, -0.0330810546875, 0.18408203125, 0.012603759765625, 0.41015625, -0.18896484375, 0.14501953125, -0.03192138671875, -0.47216796875, -0.32666015625, 0.2421875, -0.41943359375, -0.28759765625, 0.08160400390625, -0.394775390625, -0.1229248046875, -0.10107421875, -0.0306396484375, 0.927734375, 0.05084228515625, -0.313232421875, -0.0841064453125, 0.1600341796875, -0.005218505859375, -0.037841796875, 0.274658203125, -0.11383056640625, 0.2861328125, 0.58935546875, -0.45849609375, 0.2265625, -0.310546875, -0.75390625, 0.1561279296875, 0.27294921875, 0.069091796875, 0.00152587890625, 0.0545654296875, 0.15185546875, 0.05712890625, -0.202880859375, 0.420654296875, 0.67138671875]}, {"frame_number": 2, "vector": [-0.1326904296875, 0.322998046875, 0.0048828125, -0.172607421875, -0.6123046875, 0.296142578125, -0.00244140625, -0.469482421875, -0.127685546875, -0.002777099609375, -0.160888671875, 0.031982421875, 0.32958984375, 0.162109375, -0.0198974609375, 0.276611328125, 0.13525390625, -0.134521484375, -0.0689697265625, -0.44677734375, -0.51953125, -6.25, -0.3349609375, -0.69384765625, -0.00677490234375, -0.10528564453125, -0.1962890625, -0.7666015625, 0.461669921875, -0.1844482421875, -0.9453125, -0.6318359375, -0.79541015625, 0.037841796875, -0.068115234375, 0.0791015625, 0.2734375, -0.6474609375, -0.05780029296875, -0.457763671875, -0.06256103515625, 0.24951171875, 0.233154296875, 0.0330810546875, 0.9365234375, -0.343017578125, -0.415283203125, -0.218505859375, -0.5517578125, -0.135986328125, 0.04998779296875, -0.260009765625, 0.09033203125, 0.486083984375, 0.22607421875, 0.05120849609375, 0.36328125, -0.2841796875, 0.27978515625, -0.0087890625, -0.09912109375, 0.2086181640625, 0.306640625, 0.185302734375, -0.24267578125, -0.32080078125, -0.496337890625, -0.062744140625, 0.153076171875, 0.116943359375, 0.958984375, -0.1065673828125, 0.02685546875, -0.018798828125, -0.453125, 0.8623046875, -0.3369140625, -0.4384765625, 0.183837890625, -0.73828125, 1.18359375, 0.328125, 0.61376953125, 0.53466796875, -0.275390625, 0.0438232421875, 0.1988525390625, 0.01531982421875, 0.165771484375, -0.5673828125, -0.1846923828125, -0.12176513671875, 0.06640625, -0.13134765625, 0.274169921875, 0.324462890625, 0.080322265625, 0.5751953125, 0.04949951171875, -0.67822265625, 0.71484375, 0.300537109375, -0.3701171875, -0.62158203125, -0.0216522216796875, -0.1365966796875, 0.056396484375, -0.16064453125, -0.1630859375, -0.04150390625, -0.0758056640625, 0.1708984375, 0.0887451171875, -0.464111328125, 0.1961669921875, -0.5859375, -0.234619140625, -0.1639404296875, 0.346923828125, 0.5283203125, 0.03955078125, -0.26123046875, -0.0679931640625, -0.69140625, -0.1104736328125, 0.1927490234375, -0.14794921875, 0.373046875, 0.199951171875, 0.322998046875, -0.0064697265625, -0.388427734375, 0.0526123046875, 0.05377197265625, 0.5634765625, -0.035400390625, 0.0478515625, 0.7568359375, 0.609375, 0.031982421875, 0.0750732421875, 0.291015625, 0.16015625, -0.0745849609375, -0.263427734375, 0.080322265625, -0.001983642578125, -0.357666015625, 0.200439453125, 1.02734375, -0.156494140625, -1.5458984375, 0.06005859375, -0.4248046875, 0.446533203125, -0.2147216796875, -0.2900390625, 0.90966796875, 0.29638671875, -0.41552734375, -0.2255859375, 0.283447265625, 0.47900390625, 0.322509765625, -0.34521484375, -0.28369140625, -0.0302734375, 0.712890625, 0.0238037109375, 0.10400390625, 0.163330078125, 0.406494140625, -0.2059326171875, 0.6728515625, -0.236083984375, 1.0166015625, -1.109375, 0.0869140625, 0.0029296875, -0.093505859375, 0.041748046875, 0.209228515625, -0.189453125, 0.021514892578125, 0.0618896484375, 0.0396728515625, -0.00390625, 0.176025390625, -0.115234375, 0.193115234375, 0.1533203125, 0.185791015625, 0.44482421875, -0.3427734375, 0.5107421875, -0.2198486328125, 0.44091796875, -0.2607421875, -0.35888671875, 1.1259765625, 0.48095703125, -0.0458984375, -0.22607421875, 0.079345703125, -0.03857421875, 0.07080078125, 0.65869140625, 0.02569580078125, -0.662109375, -0.146240234375, 0.201904296875, -0.07666015625, 0.27197265625, -0.83251953125, -0.51953125, -0.5126953125, 0.50244140625, 1.0390625, 0.03411865234375, -0.426513671875, -0.4208984375, -0.55712890625, -0.164306640625, -0.57373046875, -0.16796875, 0.212646484375, -0.10760498046875, -0.25732421875, 0.748046875, 0.00994873046875, -0.0966796875, 0.29541015625, 0.03857421875, 0.260498046875, -0.31591796875, -0.5888671875, 0.7783203125, -0.059814453125, -0.93359375, -0.111328125, 0.7197265625, 0.2314453125, 0.0567626953125, 0.10205078125, -0.1822509765625, 1.0224609375, -0.1824951171875, 0.0260009765625, -0.444580078125, -0.11865234375, 0.00830078125, -0.04736328125, 0.30029296875, -0.013671875, -0.7021484375, -0.1131591796875, -0.176513671875, -0.02392578125, 0.490234375, 0.480712890625, -0.01495361328125, -0.17529296875, -0.097412109375, -0.00146484375, -0.265380859375, 0.0389404296875, -0.822265625, -0.2548828125, 0.56884765625, -0.4853515625, 0.05810546875, 0.397216796875, 0.364990234375, 0.25830078125, -0.11834716796875, -0.6650390625, -0.07855224609375, 0.11065673828125, -0.348388671875, -0.4130859375, -0.352783203125, 0.65771484375, -0.39453125, 0.14111328125, 0.37744140625, 0.371826171875, -0.145263671875, 0.5078125, -0.568359375, 1.1259765625, -0.02020263671875, 0.0653076171875, 0.0535888671875, -0.40576171875, 0.09722900390625, 0.20263671875, 0.15576171875, 0.40966796875, 0.62841796875, -0.279541015625, -0.024322509765625, 0.74072265625, -0.00321197509765625, -0.051025390625, -0.340576171875, 0.36572265625, 0.375732421875, -0.6845703125, 0.1251220703125, -0.06396484375, 0.35986328125, -0.373779296875, -0.025634765625, -0.279296875, -0.33251953125, 0.3935546875, -0.07403564453125, 0.0445556640625, -0.26025390625, 0.1904296875, -0.90966796875, 0.1395263671875, 0.083740234375, -0.47314453125, -0.87060546875, 0.6513671875, -0.42333984375, 0.321044921875, -0.61083984375, -0.04119873046875, -0.5537109375, -0.0416259765625, -0.1680908203125, -0.434326171875, 0.1917724609375, 0.23388671875, 0.084228515625, -0.5751953125, 0.03729248046875, 0.001708984375, -0.689453125, 0.06561279296875, -0.267578125, -0.7490234375, -0.49267578125, 0.232421875, -0.48095703125, 0.376953125, 0.218994140625, 0.033203125, -0.35693359375, -0.309814453125, 0.26025390625, 0.14453125, 0.43017578125, 0.331787109375, 0.260009765625, -0.1685791015625, -0.28466796875, -0.184814453125, -0.1513671875, 0.05322265625, 0.5146484375, 0.311279296875, -0.60302734375, -0.484130859375, -0.025146484375, 0.22314453125, 0.0882568359375, -0.0751953125, -0.06591796875, 0.269287109375, -0.179443359375, 0.30419921875, 0.310791015625, -0.060791015625, 0.40283203125, -0.261474609375, 0.1669921875, -0.52392578125, -0.0220947265625, -0.365966796875, 0.097900390625, 0.6728515625, 0.225830078125, -0.029541015625, -0.02227783203125, -0.125732421875, -0.2646484375, 0.5029296875, -0.64404296875, -0.22998046875, 0.292236328125, 0.333251953125, 1.0654296875, 0.6767578125, -0.119873046875, -0.6083984375, -0.0311279296875, -0.43994140625, 0.2470703125, 0.1600341796875, -0.1171875, -0.4384765625, -0.19384765625, 0.00732421875, -0.373046875, 0.0657958984375, 0.369140625, -0.0240478515625, 0.214111328125, -0.84228515625, -0.7314453125, -0.32421875, -0.40234375, 0.5517578125, 0.697265625, -0.2822265625, 1.05859375, -0.2396240234375, 0.11444091796875, -0.03912353515625, -0.2294921875, -0.1180419921875, -0.0302734375, -0.5068359375, -1.373046875, -0.10638427734375, 0.1163330078125, 0.192626953125, 0.07568359375, -6.265625, -0.416015625, -0.056640625, -0.71533203125, -0.1593017578125, -0.1318359375, -1.197265625, -0.63427734375, -0.26025390625, 0.0023193359375, 0.6162109375, 0.697265625, -0.23095703125, -0.047821044921875, 0.20068359375, 0.24560546875, 0.0301513671875, -0.0723876953125, 0.1513671875, 0.010009765625, 0.188720703125, -0.12359619140625, 0.0806884765625, 0.38134765625, -0.2958984375, -0.1905517578125, 0.397705078125, -0.65869140625, -0.60205078125, -0.09375, 0.184326171875, -0.27978515625, -0.3603515625, -0.443359375, -0.0550537109375, 0.484375, 0.771484375, -0.3076171875, -0.40673828125, 0.22705078125, -1.01171875, 0.30908203125, -0.0413818359375, -0.015289306640625, 0.246337890625, 0.275390625, -0.084228515625, -0.72021484375, -0.7744140625, 0.12261962890625, 0.0791015625, -1.107421875, -0.0361328125, 0.35888671875, -0.03564453125, 0.2216796875, -0.09783935546875, 0.14013671875, 0.159912109375, 0.244873046875, 0.5830078125, 0.3486328125, 0.40234375, -0.4716796875, 0.025390625, 0.06256103515625, -0.00238037109375, 0.087646484375, 0.08203125, -0.0806884765625, 0.013641357421875, 0.66064453125, -0.1484375, -0.2294921875, -0.119140625, -0.0433349609375, 0.1513671875, -0.16943359375, 0.06884765625, 0.2257080078125, 0.0435791015625, 0.1800537109375, 0.380859375, -0.18408203125, -0.17529296875, -0.107177734375, -0.3291015625, -0.5703125, -0.1336669921875, 0.7421875, 0.188232421875, -0.16259765625, 0.0679931640625, 0.20947265625, 1.322265625, 0.17724609375, 0.141357421875, 0.1746826171875, 0.0908203125, -0.3349609375, -0.5224609375, 0.01806640625, -0.059814453125, 0.39306640625, 0.32861328125, 0.0703125, -0.04461669921875, 0.44189453125, -0.105712890625, 0.748046875, 0.1492919921875, 0.353271484375, 0.47900390625, -0.2164306640625, 0.4716796875, -1.5244140625, -0.36376953125, -0.214111328125, 0.148681640625, -0.009765625, -0.32666015625, 0.783203125, 0.28369140625, -0.087646484375, 0.078369140625, 0.31884765625, -0.0657958984375, -0.239990234375, -0.89111328125, -0.452392578125, -0.273681640625, -0.43017578125, -0.96337890625, 0.266357421875, 0.28515625, 0.55419921875, -0.3251953125, -0.07562255859375, 0.0794677734375, -0.075927734375, 0.374755859375, -0.21875, 0.094482421875, 0.1982421875, 0.302001953125, -0.37255859375, 0.212646484375, -0.299560546875, -0.46923828125, -0.28271484375, 0.087646484375, -0.16259765625, -0.0980224609375, -0.0208740234375, -0.09814453125, 0.03759765625, 0.457275390625, -0.34130859375, 0.238525390625, 0.408935546875, 0.0255126953125, -0.2132568359375, 0.01666259765625, 0.0775146484375, 0.0909423828125, -0.06494140625, -0.216796875, -1.47265625, -0.4736328125, 0.1787109375, -0.05810546875, -0.3466796875, -0.350830078125, -0.20458984375, -0.19921875, -0.0906982421875, -0.28125, -0.188720703125, -0.9013671875, 0.25537109375, 0.397216796875, -0.1890869140625, -0.162353515625, -0.1104736328125, -0.57373046875, 0.241943359375, 0.2281494140625, -0.079833984375, -0.2626953125, 0.0015869140625, -0.0498046875, 0.07135009765625, 0.0643310546875, -0.236572265625, -0.025482177734375, 0.0401611328125, 0.37841796875, 0.29296875, -0.0838623046875, 0.1690673828125, 0.05938720703125, -0.3701171875, 0.7421875, 0.0623779296875, -0.108154296875, -0.336669921875, 1.34765625, 0.0947265625, -0.2724609375, 0.017333984375, -0.46875, 0.3662109375, -0.215576171875, -0.20849609375, 0.3486328125, 0.172119140625, 0.008148193359375, 0.2421875, 0.08843994140625, 0.2188720703125, -0.343017578125, 0.351806640625, 0.22900390625, -0.669921875, 0.162353515625, 0.282958984375, -0.1024169921875, -0.2744140625, -0.28125, -0.278564453125, -0.87109375, -0.005859375, -0.095703125, -0.040374755859375, -0.239501953125, 0.116455078125, -0.067138671875, 0.4580078125, -0.433349609375, -1.3828125, -0.1439208984375, -0.135986328125, 0.21533203125, 0.2255859375, -0.060302734375, 0.354736328125, 0.189208984375, -0.182373046875, 0.25927734375, -0.01470947265625, 0.1451416015625, 0.418212890625, 0.02783203125, 0.12646484375, -0.17138671875, 0.3046875, 0.1846923828125, -0.386962890625, -0.59716796875, -0.212646484375, -0.4189453125, 0.0625, 0.25927734375, -0.24560546875, 0.8056640625, -0.0758056640625, 0.087646484375, -0.873046875, 0.02276611328125, -0.0057373046875, 0.1968994140625, -0.619140625, 0.31396484375, -0.5556640625, 0.4130859375, -0.04718017578125, 0.2310791015625, -0.11553955078125, -0.5185546875, -0.626953125, 0.09466552734375, -0.448486328125, 0.4404296875, -0.333984375, 0.93994140625, -0.30615234375, 0.05889892578125, -0.1171875, -0.02880859375, -0.435546875, 0.093017578125, 1.388671875, -0.0052490234375, 0.2490234375, -0.0863037109375, 0.341796875, -1.048828125, -0.25341796875, 0.5205078125, 0.025146484375, -0.25048828125, 0.5185546875, 0.1170654296875, 0.032470703125, 0.17822265625, 0.06671142578125, 0.213134765625, -0.110107421875, 0.44189453125, -0.5517578125, 0.041259765625, -0.611328125, 0.0517578125, 0.1083984375, -0.427734375, -0.0179443359375, -0.08184814453125, -0.110595703125, 0.10546875, -0.921875, -0.24951171875, 0.28857421875, -0.150390625, 0.22412109375, 0.01971435546875, -0.0618896484375, -0.849609375, 0.03271484375, 0.316650390625, -0.00091552734375, 0.0478515625, -0.2548828125, -0.53369140625, 0.60595703125, -0.425537109375, 0.00787353515625, -0.5986328125, -0.0093994140625, -0.3681640625, 0.147705078125, 0.0126953125, 0.050048828125, 0.07122802734375, 0.1865234375, -2.279296875, 0.327392578125, -0.191650390625, 0.240234375, 0.623046875, -0.047119140625, -0.33642578125, -0.21240234375, 0.16845703125, -0.1246337890625, 0.206787109375, -0.003173828125, 0.08251953125, 0.5126953125, 0.460205078125, 0.334228515625, -0.025146484375, 0.25, 0.6337890625, -0.2685546875, 0.38671875, 0.0970458984375, 0.1070556640625, -0.266357421875, 0.162841796875, 0.25634765625, 0.28759765625, -0.740234375, 0.552734375, 0.0875244140625, -0.161865234375, 0.0103759765625, -0.5302734375, -0.0037841796875, -0.026611328125, 0.334716796875, -0.257568359375, -0.01263427734375, -0.202392578125, -0.200439453125, 0.41650390625, -0.0589599609375, -0.28369140625, -0.6455078125, -0.424560546875, 0.3125, 0.0797119140625, -0.1600341796875, 0.32861328125, -0.132080078125, 0.323974609375, -0.3935546875, -0.330078125, 0.1719970703125, -0.244873046875, -0.50048828125, -0.65869140625, -0.642578125, -0.03155517578125, 0.79638671875, 0.5654296875, -0.26123046875, -0.010986328125, 0.66455078125, -0.6533203125, 0.254638671875, -0.65380859375, -0.7197265625, -0.47998046875, 0.103515625, 0.09613037109375, -0.368896484375, -0.3076171875, 0.1748046875, -0.1591796875, 0.16015625, 0.4453125, 0.142822265625, 0.1373291015625, -0.0221099853515625, -0.50048828125, 0.079833984375, -0.0693359375, -0.08447265625, 0.7802734375, 0.118408203125, -0.3203125, 0.4931640625, -0.65234375, 0.4755859375, -0.386474609375, 0.56591796875, -0.1295166015625, 0.080810546875, 0.1221923828125, -0.9990234375, -0.131591796875, 0.06890869140625, 0.191162109375, -0.26513671875, -0.49169921875, 0.0328369140625, 1.0869140625, 0.409912109375, -0.72021484375, -0.25244140625, 0.77099609375, -0.178955078125, -0.198486328125, -0.140869140625, -0.361328125, 0.189208984375, -0.1832275390625, -0.46484375, 0.445556640625, 0.433837890625, -0.36865234375, -0.5498046875, 0.081787109375, 0.0784912109375, -0.136962890625, 0.14404296875, -0.2088623046875, -0.09442138671875, 0.251953125, 0.27001953125, 0.38916015625, 0.364501953125, 0.2021484375, -0.12744140625, 0.20361328125, -0.04443359375, 0.36767578125, 0.36181640625, 0.48974609375, -0.260986328125, -0.08709716796875, 0.1220703125, -0.2998046875, -0.01220703125, -0.1585693359375, 0.371826171875, 0.358154296875, -0.31298828125, 0.438232421875, -2.1171875, 0.5048828125, 0.40478515625, 0.153076171875, -0.56640625, -0.1424560546875, 0.06787109375, -0.152099609375, 0.0992431640625, 0.19482421875, 0.3505859375, -0.093017578125, 0.75390625, -0.380126953125, 0.083251953125, 0.7021484375, 0.264404296875, 0.24755859375, -0.232421875, -0.37841796875, 0.1844482421875, 0.55126953125, 0.1556396484375, -0.066162109375, -0.234130859375, 0.22607421875, 0.1134033203125, 0.96484375, 0.28515625, 0.322265625, -0.233154296875, 0.30712890625, 0.1181640625, 0.02825927734375, 0.6865234375, -0.6923828125, -0.2939453125, 0.5673828125, -0.6064453125, 0.07110595703125, 0.755859375, -0.47265625, -0.2666015625, -0.15283203125, -0.39208984375, -0.219970703125, 0.1055908203125, -0.9111328125, -0.1231689453125, 0.1513671875, 0.0533447265625, 0.6103515625, -0.25732421875, 0.4111328125, -0.296875, -0.2322998046875, -0.1591796875, -0.074462890625, 0.084228515625, -0.1417236328125, 0.1593017578125, 0.0660400390625, 0.12353515625, 0.357666015625, -0.16015625, -0.181884765625, -0.364501953125, 0.00804901123046875, 0.490234375, 1.2490234375, -0.1649169921875, -0.3017578125, -0.56005859375, 0.54541015625, 0.37939453125, 0.033935546875, 0.3447265625, -0.27734375, -1.3203125, -0.146240234375, 0.057373046875, -0.60888671875, 0.338134765625, 0.64892578125, -0.06341552734375, 0.196533203125, -0.1416015625, 0.06298828125, 0.13134765625, -0.015777587890625, -0.484375, 0.2890625, -0.1326904296875, -0.4755859375, 0.189208984375, -0.135986328125, -0.31787109375, -0.1817626953125, -0.206298828125, -0.2763671875, -0.406982421875, 0.033203125, 0.01171875, -0.3681640625, -0.49560546875, -0.126708984375, -0.47900390625, -0.1085205078125, 0.21484375, -0.04833984375, -0.372314453125, -0.24267578125, -0.162353515625, -0.2296142578125, 0.180908203125, 0.5126953125, 0.15234375, 0.224609375, -0.73583984375, -0.324951171875, -0.525390625, -0.2109375, 0.0106201171875, 0.1756591796875, 0.1656494140625, 0.085693359375, 0.423828125, -0.016143798828125, 0.1778564453125, 0.01104736328125, -0.169677734375, 0.43310546875, 0.211181640625, 0.57275390625, -0.16748046875, 0.0498046875, -0.7734375, -0.1229248046875, 0.0462646484375, 0.1317138671875, -0.2391357421875, -0.11297607421875, -0.529296875, 0.33935546875, -1.453125, 0.0941162109375, -0.127197265625, 1.224609375, -0.197509765625, 0.391357421875, 0.385498046875, 0.40185546875, -0.169921875, -0.08251953125, 0.276611328125, -0.05859375, 0.39697265625, 0.587890625, -0.58544921875, -0.2666015625, -0.094970703125, -0.1944580078125, -0.0625, 0.138427734375, 0.46826171875, 0.0260009765625, -0.583984375, 0.1883544921875, -0.186767578125, -0.0816650390625, -0.7724609375, 0.92138671875, 0.008056640625, 0.028076171875, -0.029541015625, -0.6611328125, 0.434814453125, 0.2093505859375, 0.4951171875, -0.06591796875, -0.63818359375, 0.69189453125, 0.25439453125, -0.1248779296875, 1.09375, 0.08447265625, -0.0882568359375, -0.76953125, -0.256591796875, -0.27685546875, -0.0257568359375, 0.16015625, 0.02581787109375, 0.4033203125, -0.188720703125, 0.129150390625, -0.012451171875, -0.48876953125, -0.320068359375, 0.250244140625, -0.40673828125, -0.30419921875, 0.08709716796875, -0.392578125, -0.13037109375, -0.0916748046875, -0.03961181640625, 0.91552734375, 0.02349853515625, -0.31103515625, -0.11083984375, 0.158935546875, 0.022552490234375, -0.029541015625, 0.292236328125, -0.11016845703125, 0.24755859375, 0.59521484375, -0.451171875, 0.213134765625, -0.2822265625, -0.775390625, 0.175048828125, 0.27685546875, 0.085205078125, 0.00604248046875, 0.0679931640625, 0.152587890625, 0.052001953125, -0.224609375, 0.42626953125, 0.69970703125]}, {"frame_number": 3, "vector": [-0.1278076171875, 0.324462890625, -0.0081787109375, -0.1845703125, -0.595703125, 0.301513671875, 0.0015869140625, -0.47314453125, -0.1112060546875, -0.0042877197265625, -0.162109375, 0.0400390625, 0.31689453125, 0.176513671875, -0.0274658203125, 0.30615234375, 0.127685546875, -0.11083984375, -0.06884765625, -0.458984375, -0.537109375, -6.2421875, -0.3505859375, -0.69140625, -0.0189208984375, -0.089599609375, -0.193359375, -0.7626953125, 0.47802734375, -0.1859130859375, -0.94384765625, -0.615234375, -0.79296875, 0.04486083984375, -0.0665283203125, 0.06689453125, 0.2861328125, -0.6298828125, -0.06744384765625, -0.440673828125, -0.077880859375, 0.22314453125, 0.2232666015625, 0.0316162109375, 0.9404296875, -0.33154296875, -0.409423828125, -0.2265625, -0.5712890625, -0.13818359375, 0.04937744140625, -0.245849609375, 0.07623291015625, 0.487548828125, 0.231689453125, 0.0640869140625, 0.38134765625, -0.290771484375, 0.2626953125, -0.03448486328125, -0.09619140625, 0.199462890625, 0.32568359375, 0.188232421875, -0.2666015625, -0.326171875, -0.5107421875, -0.021728515625, 0.1483154296875, 0.091064453125, 0.9462890625, -0.10791015625, 0.035400390625, -0.032470703125, -0.452392578125, 0.8662109375, -0.33642578125, -0.44482421875, 0.188232421875, -0.69677734375, 1.154296875, 0.332275390625, 0.62255859375, 0.5126953125, -0.26416015625, 0.049560546875, 0.185546875, 0.03369140625, 0.16015625, -0.5732421875, -0.186279296875, -0.1263427734375, 0.0849609375, -0.1396484375, 0.26611328125, 0.32861328125, 0.07373046875, 0.5673828125, 0.0771484375, -0.66845703125, 0.728515625, 0.29296875, -0.3984375, -0.626953125, -0.0352783203125, -0.1324462890625, 0.0665283203125, -0.158935546875, -0.1864013671875, -0.024658203125, -0.09637451171875, 0.171142578125, 0.09747314453125, -0.466064453125, 0.1866455078125, -0.5849609375, -0.219482421875, -0.1624755859375, 0.345703125, 0.521484375, 0.036376953125, -0.228271484375, -0.0596923828125, -0.67041015625, -0.1302490234375, 0.1832275390625, -0.1602783203125, 0.35595703125, 0.18212890625, 0.341796875, 0.0103759765625, -0.40869140625, 0.041748046875, 0.02581787109375, 0.54541015625, -0.0120849609375, 0.0521240234375, 0.744140625, 0.59521484375, 0.020751953125, 0.0513916015625, 0.30712890625, 0.1541748046875, -0.0843505859375, -0.259521484375, 0.082275390625, -0.013824462890625, -0.355224609375, 0.1888427734375, 1.01171875, -0.149658203125, -1.5244140625, 0.086181640625, -0.40625, 0.433349609375, -0.1876220703125, -0.26806640625, 0.9052734375, 0.2939453125, -0.42041015625, -0.2432861328125, 0.288818359375, 0.46875, 0.3203125, -0.33203125, -0.276123046875, -0.03759765625, 0.69287109375, 0.0316162109375, 0.11083984375, 0.1353759765625, 0.386962890625, -0.163330078125, 0.693359375, -0.25390625, 1.025390625, -1.115234375, 0.0882568359375, -0.0067138671875, -0.088623046875, 0.0135498046875, 0.20458984375, -0.176513671875, 0.0289306640625, 0.03369140625, 0.02178955078125, 0.010986328125, 0.18359375, -0.1009521484375, 0.173828125, 0.1748046875, 0.1875, 0.43701171875, -0.349609375, 0.5478515625, -0.2203369140625, 0.4755859375, -0.24462890625, -0.359375, 1.1728515625, 0.46435546875, -0.03271484375, -0.2060546875, 0.0830078125, -0.03302001953125, 0.0662841796875, 0.64306640625, 0.025726318359375, -0.671875, -0.1611328125, 0.1810302734375, -0.046630859375, 0.283203125, -0.8330078125, -0.51953125, -0.529296875, 0.5, 1.0537109375, 0.0115203857421875, -0.42626953125, -0.4404296875, -0.57958984375, -0.1827392578125, -0.59130859375, -0.1767578125, 0.184814453125, -0.129150390625, -0.2373046875, 0.73583984375, -0.00604248046875, -0.0941162109375, 0.28662109375, 0.06005859375, 0.28125, -0.33447265625, -0.5849609375, 0.7861328125, -0.0535888671875, -0.93408203125, -0.11572265625, 0.70263671875, 0.232666015625, 0.0506591796875, 0.07861328125, -0.1695556640625, 1.0234375, -0.1822509765625, 0.0115966796875, -0.440673828125, -0.116943359375, 0.023193359375, -0.04791259765625, 0.28173828125, 0.005615234375, -0.7080078125, -0.104736328125, -0.1630859375, -0.0125732421875, 0.483154296875, 0.475830078125, -0.03070068359375, -0.1767578125, -0.110107421875, 0.0118408203125, -0.2392578125, 0.03948974609375, -0.85888671875, -0.251220703125, 0.580078125, -0.46923828125, 0.0760498046875, 0.3818359375, 0.377197265625, 0.23974609375, -0.1221923828125, -0.67578125, -0.0989990234375, 0.1275634765625, -0.3486328125, -0.432373046875, -0.34912109375, 0.67626953125, -0.410400390625, 0.142822265625, 0.37255859375, 0.358154296875, -0.1563720703125, 0.52587890625, -0.56591796875, 1.134765625, 0.00506591796875, 0.06866455078125, 0.0513916015625, -0.388427734375, 0.10626220703125, 0.19970703125, 0.151611328125, 0.41552734375, 0.6416015625, -0.267333984375, -0.037353515625, 0.7626953125, -0.01824951171875, -0.056640625, -0.335205078125, 0.354248046875, 0.34765625, -0.68408203125, 0.123046875, -0.060791015625, 0.38671875, -0.36669921875, -0.03662109375, -0.30712890625, -0.32080078125, 0.401611328125, -0.06298828125, 0.04248046875, -0.241455078125, 0.215087890625, -0.90087890625, 0.1214599609375, 0.1063232421875, -0.4736328125, -0.88525390625, 0.669921875, -0.4033203125, 0.317626953125, -0.63232421875, -0.02606201171875, -0.5361328125, -0.0389404296875, -0.181640625, -0.4521484375, 0.200927734375, 0.2451171875, 0.067626953125, -0.58984375, 0.03948974609375, 0.01416015625, -0.697265625, 0.063720703125, -0.281494140625, -0.7109375, -0.5068359375, 0.2333984375, -0.4931640625, 0.3740234375, 0.2432861328125, 0.025634765625, -0.34033203125, -0.272705078125, 0.270263671875, 0.12841796875, 0.4248046875, 0.34326171875, 0.2486572265625, -0.1602783203125, -0.303955078125, -0.148193359375, -0.1806640625, 0.056884765625, 0.5126953125, 0.326904296875, -0.611328125, -0.4853515625, -0.031982421875, 0.22998046875, 0.09405517578125, -0.07427978515625, -0.05242919921875, 0.254150390625, -0.196044921875, 0.275390625, 0.278564453125, -0.054931640625, 0.4150390625, -0.25927734375, 0.166015625, -0.5361328125, -0.0599365234375, -0.368408203125, 0.12109375, 0.6875, 0.251220703125, -0.05078125, -0.02813720703125, -0.137939453125, -0.25439453125, 0.498779296875, -0.66796875, -0.221435546875, 0.30322265625, 0.326171875, 1.046875, 0.6708984375, -0.138916015625, -0.6142578125, -0.039306640625, -0.4580078125, 0.25927734375, 0.1689453125, -0.12109375, -0.4501953125, -0.1995849609375, 0.014892578125, -0.357421875, 0.0792236328125, 0.37841796875, -0.019287109375, 0.201416015625, -0.8310546875, -0.69921875, -0.333251953125, -0.40673828125, 0.56005859375, 0.6875, -0.2841796875, 1.0625, -0.229736328125, 0.1112060546875, -0.050445556640625, -0.231201171875, -0.126220703125, -0.0242919921875, -0.501953125, -1.39453125, -0.1156005859375, 0.1029052734375, 0.2054443359375, 0.076171875, -6.25, -0.39111328125, -0.083984375, -0.7333984375, -0.168212890625, -0.1280517578125, -1.185546875, -0.62890625, -0.241943359375, 0.023681640625, 0.61572265625, 0.6884765625, -0.232421875, -0.0738525390625, 0.20458984375, 0.22021484375, 0.03076171875, -0.0736083984375, 0.1689453125, 0.007568359375, 0.187744140625, -0.131591796875, 0.0657958984375, 0.408203125, -0.305908203125, -0.1888427734375, 0.4306640625, -0.64990234375, -0.61328125, -0.077392578125, 0.1943359375, -0.281005859375, -0.354248046875, -0.45458984375, -0.0364990234375, 0.47705078125, 0.76025390625, -0.30419921875, -0.41162109375, 0.2144775390625, -1.015625, 0.306396484375, -0.0452880859375, -0.0260009765625, 0.2177734375, 0.2841796875, -0.0950927734375, -0.7431640625, -0.7900390625, 0.1552734375, 0.0499267578125, -1.099609375, -0.01806640625, 0.3681640625, -0.0438232421875, 0.208251953125, -0.10308837890625, 0.1611328125, 0.1611328125, 0.241943359375, 0.60205078125, 0.326171875, 0.396484375, -0.48779296875, 0.036865234375, 0.06793212890625, 0.025634765625, 0.0888671875, 0.0753173828125, -0.0755615234375, 0.0439453125, 0.6904296875, -0.166015625, -0.23046875, -0.09326171875, -0.0391845703125, 0.1243896484375, -0.169677734375, 0.06396484375, 0.20703125, 0.0439453125, 0.190673828125, 0.373046875, -0.18896484375, -0.1478271484375, -0.10400390625, -0.336669921875, -0.57861328125, -0.1337890625, 0.75048828125, 0.196044921875, -0.154541015625, 0.06793212890625, 0.232666015625, 1.3046875, 0.16845703125, 0.150634765625, 0.1644287109375, 0.096923828125, -0.33837890625, -0.50732421875, 0.00732421875, -0.067138671875, 0.365234375, 0.342041015625, 0.07373046875, -0.0279541015625, 0.4375, -0.1025390625, 0.732421875, 0.1402587890625, 0.34619140625, 0.48095703125, -0.2098388671875, 0.49072265625, -1.5, -0.36572265625, -0.22900390625, 0.147705078125, 0.0081787109375, -0.342529296875, 0.779296875, 0.29345703125, -0.0941162109375, 0.08349609375, 0.30029296875, -0.0830078125, -0.233154296875, -0.89013671875, -0.441650390625, -0.271240234375, -0.425537109375, -0.96044921875, 0.25244140625, 0.28466796875, 0.5751953125, -0.32177734375, -0.0704345703125, 0.06768798828125, -0.0830078125, 0.384765625, -0.208984375, 0.11932373046875, 0.194091796875, 0.303955078125, -0.3583984375, 0.201904296875, -0.333984375, -0.468505859375, -0.3037109375, 0.0926513671875, -0.16455078125, -0.0828857421875, 0.0003662109375, -0.08538818359375, 0.0322265625, 0.464111328125, -0.347900390625, 0.2274169921875, 0.39111328125, 0.0283203125, -0.21826171875, -0.00469970703125, 0.08563232421875, 0.0567626953125, -0.0645751953125, -0.2203369140625, -1.4638671875, -0.454833984375, 0.1690673828125, -0.03546142578125, -0.3447265625, -0.36962890625, -0.215576171875, -0.1846923828125, -0.0628662109375, -0.29296875, -0.1890869140625, -0.88232421875, 0.25, 0.41064453125, -0.1837158203125, -0.158447265625, -0.109130859375, -0.58984375, 0.255126953125, 0.2427978515625, -0.057861328125, -0.25439453125, -0.0018310546875, -0.0859375, 0.088623046875, 0.0941162109375, -0.248779296875, -0.0176239013671875, 0.04425048828125, 0.396484375, 0.287353515625, -0.0596923828125, 0.1845703125, 0.06671142578125, -0.3720703125, 0.7333984375, 0.0850830078125, -0.117919921875, -0.322509765625, 1.322265625, 0.070068359375, -0.275146484375, 0.0230712890625, -0.46435546875, 0.36767578125, -0.2001953125, -0.201171875, 0.353515625, 0.160888671875, 0.0111083984375, 0.246337890625, 0.0841064453125, 0.2369384765625, -0.34814453125, 0.3486328125, 0.240478515625, -0.66259765625, 0.144287109375, 0.281494140625, -0.0968017578125, -0.287109375, -0.26953125, -0.289794921875, -0.8876953125, -0.00732421875, -0.06884765625, -0.0526123046875, -0.236083984375, 0.123291015625, -0.08251953125, 0.443359375, -0.4287109375, -1.37890625, -0.1253662109375, -0.13623046875, 0.21630859375, 0.2098388671875, -0.06744384765625, 0.34033203125, 0.172607421875, -0.167236328125, 0.2646484375, -0.01922607421875, 0.165771484375, 0.423828125, 0.033203125, 0.1109619140625, -0.159912109375, 0.301025390625, 0.1654052734375, -0.3935546875, -0.5966796875, -0.2115478515625, -0.41357421875, 0.038818359375, 0.25146484375, -0.26416015625, 0.806640625, -0.0777587890625, 0.0838623046875, -0.86474609375, 0.02056884765625, -0.01025390625, 0.1966552734375, -0.5986328125, 0.30419921875, -0.5498046875, 0.40185546875, -0.047119140625, 0.22802734375, -0.10595703125, -0.53564453125, -0.61962890625, 0.11651611328125, -0.420654296875, 0.4306640625, -0.330810546875, 0.93994140625, -0.31005859375, 0.04638671875, -0.126708984375, -0.0216064453125, -0.4345703125, 0.082275390625, 1.349609375, 0.00860595703125, 0.249267578125, -0.0887451171875, 0.35400390625, -1.064453125, -0.248046875, 0.509765625, -0.0079345703125, -0.279541015625, 0.51025390625, 0.135009765625, 0.0418701171875, 0.16650390625, 0.08489990234375, 0.1953125, -0.129150390625, 0.46630859375, -0.5380859375, 0.02734375, -0.62255859375, 0.0526123046875, 0.1107177734375, -0.44873046875, 0.00579833984375, -0.069091796875, -0.12646484375, 0.12060546875, -0.9208984375, -0.2705078125, 0.298583984375, -0.14599609375, 0.228515625, 0.022705078125, -0.07861328125, -0.83203125, 0.0238037109375, 0.312255859375, 0.01995849609375, 0.056549072265625, -0.256103515625, -0.52099609375, 0.615234375, -0.412841796875, -0.0035400390625, -0.603515625, -0.0135498046875, -0.34130859375, 0.149169921875, 0.01904296875, 0.044677734375, 0.0736083984375, 0.1881103515625, -2.28125, 0.352783203125, -0.21923828125, 0.258056640625, 0.61572265625, -0.0565185546875, -0.3310546875, -0.21435546875, 0.17626953125, -0.1278076171875, 0.19140625, -0.021728515625, 0.0679931640625, 0.49267578125, 0.4775390625, 0.32373046875, -0.016357421875, 0.2469482421875, 0.6318359375, -0.27734375, 0.40234375, 0.11553955078125, 0.1341552734375, -0.25634765625, 0.1279296875, 0.25341796875, 0.27978515625, -0.75439453125, 0.541015625, 0.0673828125, -0.2198486328125, 0.0152587890625, -0.52197265625, -0.001953125, -0.0478515625, 0.342529296875, -0.245849609375, -0.042083740234375, -0.19873046875, -0.187255859375, 0.42626953125, -0.063720703125, -0.27294921875, -0.6533203125, -0.4228515625, 0.31103515625, 0.09326171875, -0.1710205078125, 0.328125, -0.1253662109375, 0.3125, -0.408203125, -0.3330078125, 0.1796875, -0.235595703125, -0.459716796875, -0.646484375, -0.65234375, -0.02825927734375, 0.7841796875, 0.59326171875, -0.25732421875, -0.017822265625, 0.66552734375, -0.66796875, 0.2548828125, -0.64208984375, -0.69287109375, -0.4833984375, 0.111328125, 0.08843994140625, -0.357421875, -0.34033203125, 0.1656494140625, -0.12353515625, 0.14306640625, 0.46728515625, 0.1351318359375, 0.127197265625, -0.0152130126953125, -0.48046875, 0.068603515625, -0.063232421875, -0.074462890625, 0.8056640625, 0.11181640625, -0.301513671875, 0.487548828125, -0.685546875, 0.46826171875, -0.38818359375, 0.5556640625, -0.123046875, 0.078369140625, 0.122314453125, -1.01953125, -0.133544921875, 0.05078125, 0.19580078125, -0.27734375, -0.478271484375, 0.03668212890625, 1.08203125, 0.38671875, -0.71826171875, -0.253662109375, 0.74853515625, -0.172119140625, -0.212158203125, -0.140869140625, -0.370361328125, 0.158935546875, -0.1923828125, -0.455322265625, 0.435546875, 0.434326171875, -0.361328125, -0.52392578125, 0.09466552734375, 0.091796875, -0.1337890625, 0.1436767578125, -0.1943359375, -0.06280517578125, 0.253662109375, 0.258544921875, 0.380859375, 0.380615234375, 0.20263671875, -0.1407470703125, 0.169189453125, -0.0439453125, 0.3935546875, 0.35107421875, 0.5009765625, -0.276123046875, -0.083251953125, 0.1474609375, -0.3056640625, 0.0029296875, -0.1524658203125, 0.3974609375, 0.363525390625, -0.3134765625, 0.449951171875, -2.103515625, 0.498779296875, 0.425048828125, 0.11279296875, -0.5498046875, -0.15478515625, 0.0693359375, -0.145751953125, 0.0933837890625, 0.2158203125, 0.35205078125, -0.113037109375, 0.73583984375, -0.375732421875, 0.07513427734375, 0.7060546875, 0.258544921875, 0.24267578125, -0.21044921875, -0.38818359375, 0.2037353515625, 0.5517578125, 0.1629638671875, -0.07421875, -0.2122802734375, 0.23095703125, 0.12286376953125, 0.93896484375, 0.28759765625, 0.336669921875, -0.223876953125, 0.30859375, 0.115234375, 0.026611328125, 0.6728515625, -0.6748046875, -0.27880859375, 0.56396484375, -0.61767578125, 0.0587158203125, 0.748046875, -0.48193359375, -0.273681640625, -0.13623046875, -0.38623046875, -0.234130859375, 0.103515625, -0.90380859375, -0.116455078125, 0.1636962890625, 0.05322265625, 0.60009765625, -0.2646484375, 0.42578125, -0.310791015625, -0.236083984375, -0.1788330078125, -0.080810546875, 0.070068359375, -0.102294921875, 0.1363525390625, 0.05389404296875, 0.131103515625, 0.361083984375, -0.166015625, -0.194580078125, -0.375, 0.02252197265625, 0.4599609375, 1.205078125, -0.158203125, -0.322998046875, -0.5830078125, 0.5625, 0.34228515625, 0.030517578125, 0.30224609375, -0.277099609375, -1.32421875, -0.1317138671875, 0.05078125, -0.625, 0.3671875, 0.62841796875, -0.07110595703125, 0.21875, -0.1483154296875, 0.0640869140625, 0.13720703125, -0.023406982421875, -0.4716796875, 0.31591796875, -0.1358642578125, -0.4609375, 0.1976318359375, -0.132568359375, -0.2919921875, -0.1988525390625, -0.2159423828125, -0.28466796875, -0.380615234375, 0.03466796875, 0.018798828125, -0.353759765625, -0.5029296875, -0.12939453125, -0.50537109375, -0.09759521484375, 0.201416015625, -0.082763671875, -0.370849609375, -0.2373046875, -0.179931640625, -0.2113037109375, 0.171630859375, 0.509765625, 0.149658203125, 0.209716796875, -0.73046875, -0.326416015625, -0.5302734375, -0.2205810546875, 0.0064697265625, 0.198974609375, 0.143310546875, 0.07330322265625, 0.4091796875, 0.011871337890625, 0.190185546875, 0.01800537109375, -0.1712646484375, 0.43359375, 0.211669921875, 0.57177734375, -0.21484375, 0.03955078125, -0.7861328125, -0.114501953125, 0.04052734375, 0.137939453125, -0.24560546875, -0.08740234375, -0.53125, 0.31884765625, -1.4443359375, 0.1142578125, -0.125, 1.228515625, -0.1767578125, 0.381591796875, 0.382568359375, 0.396484375, -0.1767578125, -0.0888671875, 0.273193359375, -0.0313720703125, 0.385498046875, 0.5751953125, -0.5888671875, -0.27392578125, -0.11517333984375, -0.2001953125, -0.05340576171875, 0.1455078125, 0.46142578125, 0.041107177734375, -0.59912109375, 0.1890869140625, -0.178955078125, -0.0887451171875, -0.7529296875, 0.880859375, 0.0269775390625, 0.0225830078125, -0.0352783203125, -0.6611328125, 0.4453125, 0.2130126953125, 0.5126953125, -0.05712890625, -0.6220703125, 0.658203125, 0.28076171875, -0.114501953125, 1.09765625, 0.04931640625, -0.1097412109375, -0.74267578125, -0.23681640625, -0.302978515625, -0.0191650390625, 0.15283203125, 0.0240631103515625, 0.39306640625, -0.19677734375, 0.1312255859375, -0.0185546875, -0.45361328125, -0.323974609375, 0.23681640625, -0.41650390625, -0.30859375, 0.083740234375, -0.394287109375, -0.1346435546875, -0.10455322265625, -0.05419921875, 0.94482421875, 0.0042724609375, -0.3115234375, -0.1043701171875, 0.146240234375, 0.028564453125, -0.0262451171875, 0.283447265625, -0.1126708984375, 0.270751953125, 0.5869140625, -0.44580078125, 0.2041015625, -0.29833984375, -0.77197265625, 0.1612548828125, 0.261474609375, 0.10693359375, 0.01116943359375, 0.0684814453125, 0.152587890625, 0.068115234375, -0.21142578125, 0.42578125, 0.69921875]}, {"frame_number": 4, "vector": [-0.1666259765625, 0.314208984375, 0.01025390625, -0.185546875, -0.609375, 0.3056640625, 0.01287841796875, -0.475341796875, -0.103271484375, -0.029327392578125, -0.157470703125, 0.040283203125, 0.31689453125, 0.17919921875, -0.02490234375, 0.295654296875, 0.13134765625, -0.131103515625, -0.080078125, -0.46142578125, -0.5361328125, -6.171875, -0.338623046875, -0.66796875, -0.01788330078125, -0.0640869140625, -0.2138671875, -0.77001953125, 0.48876953125, -0.19091796875, -0.9306640625, -0.63232421875, -0.79638671875, 0.05291748046875, -0.06951904296875, 0.04278564453125, 0.28271484375, -0.62841796875, -0.060791015625, -0.439697265625, -0.066650390625, 0.23583984375, 0.2332763671875, 0.0333251953125, 0.98486328125, -0.361572265625, -0.38427734375, -0.24267578125, -0.5537109375, -0.15478515625, 0.041351318359375, -0.24609375, 0.0928955078125, 0.49365234375, 0.241943359375, 0.03619384765625, 0.39599609375, -0.27099609375, 0.271240234375, -0.0174560546875, -0.113037109375, 0.1962890625, 0.31005859375, 0.182861328125, -0.2578125, -0.330078125, -0.51171875, -0.0361328125, 0.1650390625, 0.1033935546875, 0.982421875, -0.0904541015625, 0.004638671875, -0.0345458984375, -0.439208984375, 0.873046875, -0.334716796875, -0.4267578125, 0.21630859375, -0.67578125, 1.154296875, 0.3369140625, 0.63134765625, 0.53955078125, -0.265625, 0.0531005859375, 0.19140625, 0.01953125, 0.181640625, -0.5859375, -0.18408203125, -0.11871337890625, 0.0626220703125, -0.11083984375, 0.26123046875, 0.28759765625, 0.1015625, 0.5615234375, 0.08380126953125, -0.68896484375, 0.68896484375, 0.317626953125, -0.36474609375, -0.6337890625, -0.043212890625, -0.1175537109375, 0.06915283203125, -0.163818359375, -0.177734375, -0.05126953125, -0.10888671875, 0.170166015625, 0.07293701171875, -0.44384765625, 0.1778564453125, -0.5947265625, -0.229248046875, -0.16259765625, 0.3349609375, 0.52685546875, 0.03955078125, -0.25390625, -0.044921875, -0.630859375, -0.1007080078125, 0.16748046875, -0.146240234375, 0.3662109375, 0.21142578125, 0.34912109375, 0.01171875, -0.40234375, 0.0003662109375, 0.01611328125, 0.5234375, -0.0367431640625, 0.0335693359375, 0.74072265625, 0.603515625, 0.033447265625, 0.0242919921875, 0.303955078125, 0.1558837890625, -0.079345703125, -0.2724609375, 0.1102294921875, -0.0059967041015625, -0.362548828125, 0.214111328125, 1.00390625, -0.140380859375, -1.537109375, 0.099609375, -0.4072265625, 0.460693359375, -0.1959228515625, -0.2666015625, 0.91845703125, 0.298095703125, -0.405517578125, -0.2430419921875, 0.258056640625, 0.48974609375, 0.30712890625, -0.3271484375, -0.27685546875, -0.0277099609375, 0.7138671875, 0.0592041015625, 0.0872802734375, 0.14892578125, 0.382568359375, -0.1788330078125, 0.6962890625, -0.26611328125, 1.01953125, -1.111328125, 0.092529296875, 0.0238037109375, -0.068359375, 0.055908203125, 0.22314453125, -0.157470703125, 0.022979736328125, 0.031982421875, 0.02508544921875, -0.01513671875, 0.177001953125, -0.10498046875, 0.1842041015625, 0.16162109375, 0.1859130859375, 0.438232421875, -0.35009765625, 0.5263671875, -0.23876953125, 0.4853515625, -0.271240234375, -0.376220703125, 1.140625, 0.495361328125, -0.04736328125, -0.236572265625, 0.0662841796875, -0.0220947265625, 0.08563232421875, 0.66943359375, 0.0269775390625, -0.673828125, -0.152587890625, 0.1800537109375, -0.089599609375, 0.26513671875, -0.8125, -0.5029296875, -0.55224609375, 0.5126953125, 1.0703125, 0.037017822265625, -0.4326171875, -0.42822265625, -0.5732421875, -0.1943359375, -0.58447265625, -0.16748046875, 0.177001953125, -0.087646484375, -0.23046875, 0.73486328125, 0.00738525390625, -0.0970458984375, 0.2880859375, 0.0030517578125, 0.2900390625, -0.327880859375, -0.6103515625, 0.7626953125, -0.0682373046875, -0.912109375, -0.117919921875, 0.72021484375, 0.25390625, 0.064453125, 0.090087890625, -0.20068359375, 1.037109375, -0.1871337890625, -0.002685546875, -0.412353515625, -0.14697265625, 0.035888671875, -0.04168701171875, 0.29541015625, -0.015380859375, -0.69921875, -0.1234130859375, -0.1710205078125, -0.031494140625, 0.490478515625, 0.47998046875, -0.0123291015625, -0.176025390625, -0.092041015625, 0.0126953125, -0.233642578125, 0.06884765625, -0.8720703125, -0.237060546875, 0.5751953125, -0.454833984375, 0.0906982421875, 0.40234375, 0.3701171875, 0.25244140625, -0.130126953125, -0.671875, -0.093017578125, 0.123046875, -0.337646484375, -0.43603515625, -0.3359375, 0.66796875, -0.403564453125, 0.112060546875, 0.39013671875, 0.368896484375, -0.1497802734375, 0.53759765625, -0.552734375, 1.1435546875, -0.039306640625, 0.073974609375, 0.0546875, -0.3369140625, 0.09600830078125, 0.18505859375, 0.142333984375, 0.3798828125, 0.6328125, -0.2705078125, -0.0248870849609375, 0.78369140625, -0.0069580078125, -0.0703125, -0.33154296875, 0.3544921875, 0.357177734375, -0.6884765625, 0.1334228515625, -0.0697021484375, 0.357177734375, -0.366943359375, -0.023193359375, -0.3037109375, -0.3232421875, 0.41259765625, -0.07757568359375, 0.0540771484375, -0.227294921875, 0.197265625, -0.884765625, 0.14306640625, 0.1015625, -0.48779296875, -0.859375, 0.6484375, -0.402587890625, 0.2822265625, -0.6005859375, -0.02880859375, -0.5556640625, -0.053955078125, -0.155517578125, -0.448486328125, 0.208740234375, 0.2449951171875, 0.08331298828125, -0.63671875, 0.05474853515625, 0.0093994140625, -0.6923828125, 0.05029296875, -0.2724609375, -0.70751953125, -0.51171875, 0.248779296875, -0.489990234375, 0.3720703125, 0.27001953125, 0.0362548828125, -0.37109375, -0.283447265625, 0.296142578125, 0.1357421875, 0.424072265625, 0.3427734375, 0.2578125, -0.1541748046875, -0.2724609375, -0.18701171875, -0.174560546875, 0.063720703125, 0.5009765625, 0.322509765625, -0.6025390625, -0.478515625, -0.037109375, 0.233642578125, 0.06927490234375, -0.060791015625, -0.0670166015625, 0.24755859375, -0.174072265625, 0.309814453125, 0.32080078125, -0.0888671875, 0.404052734375, -0.296875, 0.15576171875, -0.560546875, -0.0858154296875, -0.3466796875, 0.1112060546875, 0.7080078125, 0.2332763671875, -0.04248046875, -0.0218505859375, -0.128173828125, -0.2479248046875, 0.5322265625, -0.6376953125, -0.22998046875, 0.283935546875, 0.302734375, 1.052734375, 0.6865234375, -0.148681640625, -0.63232421875, -0.0343017578125, -0.40478515625, 0.232666015625, 0.1865234375, -0.123046875, -0.4423828125, -0.215087890625, 0.00537109375, -0.391845703125, 0.0499267578125, 0.384033203125, -0.02197265625, 0.255615234375, -0.830078125, -0.7236328125, -0.32275390625, -0.415283203125, 0.56591796875, 0.70947265625, -0.27392578125, 1.0390625, -0.2432861328125, 0.115966796875, -0.06536865234375, -0.259765625, -0.1044921875, -0.0115966796875, -0.51171875, -1.40234375, -0.12548828125, 0.1109619140625, 0.211181640625, 0.0802001953125, -6.17578125, -0.38525390625, -0.053466796875, -0.6953125, -0.183837890625, -0.1044921875, -1.208984375, -0.62744140625, -0.25146484375, -0.003662109375, 0.59912109375, 0.67822265625, -0.2431640625, -0.0631103515625, 0.2491455078125, 0.220947265625, 0.0208740234375, -0.10205078125, 0.163330078125, -0.014892578125, 0.184814453125, -0.08636474609375, 0.0830078125, 0.416748046875, -0.29296875, -0.1868896484375, 0.418701171875, -0.65966796875, -0.6123046875, -0.10107421875, 0.182373046875, -0.28369140625, -0.3623046875, -0.47998046875, -0.0511474609375, 0.470703125, 0.75732421875, -0.302734375, -0.4091796875, 0.22607421875, -1.0244140625, 0.30078125, -0.0499267578125, -0.0281982421875, 0.2568359375, 0.2880859375, -0.076171875, -0.7353515625, -0.8046875, 0.1348876953125, 0.0318603515625, -1.125, -0.0400390625, 0.380126953125, -0.0455322265625, 0.238525390625, -0.1156005859375, 0.13037109375, 0.177490234375, 0.26025390625, 0.58154296875, 0.32958984375, 0.373046875, -0.485595703125, 0.04638671875, 0.108642578125, 0.0289306640625, 0.098876953125, 0.089111328125, -0.056396484375, 0.03607177734375, 0.6826171875, -0.174560546875, -0.263671875, -0.1181640625, -0.0482177734375, 0.17724609375, -0.163818359375, 0.076416015625, 0.2049560546875, 0.03076171875, 0.1767578125, 0.377685546875, -0.18505859375, -0.1553955078125, -0.09521484375, -0.357421875, -0.5986328125, -0.095947265625, 0.7626953125, 0.1748046875, -0.169921875, 0.05328369140625, 0.2197265625, 1.263671875, 0.1669921875, 0.146728515625, 0.1634521484375, 0.086181640625, -0.353271484375, -0.48828125, -0.00244140625, -0.04638671875, 0.389404296875, 0.360107421875, 0.0634765625, -0.0545654296875, 0.4306640625, -0.0899658203125, 0.73193359375, 0.13916015625, 0.352294921875, 0.44970703125, -0.2392578125, 0.48046875, -1.4892578125, -0.36767578125, -0.2313232421875, 0.1531982421875, 0.0174560546875, -0.336669921875, 0.751953125, 0.2890625, -0.073486328125, 0.087646484375, 0.365234375, -0.046875, -0.252685546875, -0.896484375, -0.4453125, -0.29296875, -0.400634765625, -0.91796875, 0.261474609375, 0.28125, 0.6005859375, -0.359375, -0.07769775390625, 0.09075927734375, -0.093994140625, 0.38232421875, -0.25, 0.1116943359375, 0.198486328125, 0.32763671875, -0.40478515625, 0.2197265625, -0.320556640625, -0.45556640625, -0.291015625, 0.0731201171875, -0.16162109375, -0.1075439453125, -0.017578125, -0.0670166015625, 0.0390625, 0.482421875, -0.3427734375, 0.20751953125, 0.38623046875, 0.024627685546875, -0.168701171875, 0.007598876953125, 0.087646484375, 0.0721435546875, -0.0604248046875, -0.216064453125, -1.470703125, -0.470703125, 0.184326171875, -0.04632568359375, -0.33740234375, -0.34375, -0.2294921875, -0.1712646484375, -0.058807373046875, -0.298095703125, -0.208251953125, -0.8876953125, 0.212646484375, 0.403564453125, -0.2078857421875, -0.173095703125, -0.1143798828125, -0.57861328125, 0.24951171875, 0.224609375, -0.05303955078125, -0.244140625, 0.01177978515625, -0.08251953125, 0.06048583984375, 0.0791015625, -0.27587890625, 0.0063934326171875, 0.037109375, 0.378662109375, 0.29150390625, -0.0404052734375, 0.185791015625, 0.047607421875, -0.381103515625, 0.69873046875, 0.056884765625, -0.11181640625, -0.327880859375, 1.328125, 0.09619140625, -0.246337890625, 0.042724609375, -0.48095703125, 0.370361328125, -0.203369140625, -0.197265625, 0.332275390625, 0.15966796875, 0.01202392578125, 0.248779296875, 0.10430908203125, 0.22119140625, -0.3369140625, 0.34375, 0.22216796875, -0.6708984375, 0.14404296875, 0.27392578125, -0.0938720703125, -0.2685546875, -0.25927734375, -0.2763671875, -0.87744140625, -0.0093994140625, -0.0587158203125, -0.07373046875, -0.215576171875, 0.11669921875, -0.062744140625, 0.4755859375, -0.4208984375, -1.380859375, -0.12158203125, -0.1075439453125, 0.20751953125, 0.1688232421875, -0.05706787109375, 0.3388671875, 0.1451416015625, -0.169189453125, 0.2257080078125, -0.021240234375, 0.181396484375, 0.408447265625, 0.041015625, 0.14404296875, -0.195556640625, 0.30615234375, 0.1778564453125, -0.384033203125, -0.6123046875, -0.21533203125, -0.43701171875, 0.03076171875, 0.25927734375, -0.246826171875, 0.8173828125, -0.068115234375, 0.0621337890625, -0.91259765625, 0.034423828125, -0.008056640625, 0.22509765625, -0.6103515625, 0.29052734375, -0.5439453125, 0.4111328125, -0.0372314453125, 0.2381591796875, -0.11907958984375, -0.489501953125, -0.5927734375, 0.0848388671875, -0.438720703125, 0.44873046875, -0.339599609375, 0.95068359375, -0.2880859375, 0.05047607421875, -0.123779296875, -0.02685546875, -0.447265625, 0.064697265625, 1.380859375, -0.00738525390625, 0.269775390625, -0.0853271484375, 0.359375, -1.025390625, -0.236572265625, 0.52294921875, 0.03289794921875, -0.24462890625, 0.537109375, 0.130859375, 0.0618896484375, 0.156494140625, 0.042083740234375, 0.20703125, -0.142333984375, 0.43701171875, -0.5947265625, 0.0498046875, -0.619140625, 0.0758056640625, 0.0948486328125, -0.43310546875, 0.018798828125, -0.0748291015625, -0.138916015625, 0.12744140625, -0.896484375, -0.2666015625, 0.30810546875, -0.1448974609375, 0.2369384765625, 0.03082275390625, -0.0845947265625, -0.8466796875, 0.0328369140625, 0.30419921875, 0.0311279296875, 0.0853271484375, -0.281494140625, -0.51904296875, 0.60302734375, -0.417236328125, 0.00726318359375, -0.60302734375, -0.0079345703125, -0.3486328125, 0.1541748046875, 0.02197265625, 0.085693359375, 0.07659912109375, 0.1888427734375, -2.318359375, 0.352294921875, -0.18798828125, 0.263671875, 0.63818359375, -0.042236328125, -0.3427734375, -0.22900390625, 0.2099609375, -0.1322021484375, 0.214111328125, -0.025390625, 0.0374755859375, 0.47119140625, 0.439697265625, 0.330322265625, 0.0032958984375, 0.25146484375, 0.6181640625, -0.26416015625, 0.400146484375, 0.09405517578125, 0.1085205078125, -0.2919921875, 0.097900390625, 0.275634765625, 0.27734375, -0.7197265625, 0.533203125, 0.0567626953125, -0.170654296875, -0.0064697265625, -0.50634765625, -0.015869140625, -0.0338134765625, 0.36181640625, -0.24169921875, -0.006134033203125, -0.19189453125, -0.1611328125, 0.383544921875, -0.07318115234375, -0.2744140625, -0.6337890625, -0.414306640625, 0.302734375, 0.0762939453125, -0.167724609375, 0.342041015625, -0.12359619140625, 0.32470703125, -0.43359375, -0.36669921875, 0.1553955078125, -0.236083984375, -0.46142578125, -0.64306640625, -0.64697265625, -0.03802490234375, 0.82763671875, 0.611328125, -0.27001953125, -0.0291748046875, 0.65673828125, -0.669921875, 0.2119140625, -0.69482421875, -0.71728515625, -0.46728515625, 0.13818359375, 0.08074951171875, -0.383544921875, -0.334228515625, 0.17626953125, -0.119140625, 0.15869140625, 0.471435546875, 0.1695556640625, 0.14013671875, 0.006103515625, -0.4921875, 0.0712890625, -0.1092529296875, -0.0762939453125, 0.7822265625, 0.1171875, -0.29052734375, 0.48486328125, -0.6494140625, 0.47802734375, -0.3798828125, 0.544921875, -0.1351318359375, 0.084228515625, 0.12939453125, -1.001953125, -0.1494140625, 0.0379638671875, 0.189453125, -0.300048828125, -0.462158203125, 0.0216064453125, 1.0966796875, 0.3740234375, -0.7109375, -0.2451171875, 0.759765625, -0.16015625, -0.215576171875, -0.13525390625, -0.356201171875, 0.170654296875, -0.2049560546875, -0.46435546875, 0.44970703125, 0.43603515625, -0.336669921875, -0.521484375, 0.08612060546875, 0.0765380859375, -0.12451171875, 0.118896484375, -0.263671875, -0.06317138671875, 0.26416015625, 0.243408203125, 0.3984375, 0.396484375, 0.21044921875, -0.138916015625, 0.172119140625, -0.03851318359375, 0.3759765625, 0.355224609375, 0.486328125, -0.2568359375, -0.0927734375, 0.12646484375, -0.28076171875, -0.021484375, -0.16162109375, 0.365234375, 0.404296875, -0.304443359375, 0.451171875, -2.12109375, 0.48388671875, 0.42333984375, 0.1328125, -0.5517578125, -0.154052734375, 0.0748291015625, -0.16845703125, 0.09454345703125, 0.18017578125, 0.363525390625, -0.139404296875, 0.7568359375, -0.38232421875, 0.0909423828125, 0.70849609375, 0.26220703125, 0.27001953125, -0.23876953125, -0.386962890625, 0.18798828125, 0.54052734375, 0.156982421875, -0.07489013671875, -0.2235107421875, 0.219482421875, 0.11590576171875, 0.970703125, 0.255859375, 0.336669921875, -0.238037109375, 0.3330078125, 0.103515625, 0.0479736328125, 0.70361328125, -0.7109375, -0.26806640625, 0.5478515625, -0.6005859375, 0.0986328125, 0.7392578125, -0.46728515625, -0.284912109375, -0.164306640625, -0.39501953125, -0.1953125, 0.0810546875, -0.8798828125, -0.104248046875, 0.1256103515625, 0.062744140625, 0.5966796875, -0.29248046875, 0.434814453125, -0.305908203125, -0.236572265625, -0.172119140625, -0.086669921875, 0.06396484375, -0.1446533203125, 0.1695556640625, 0.05682373046875, 0.123046875, 0.3525390625, -0.17529296875, -0.168212890625, -0.375732421875, 0.014862060546875, 0.4892578125, 1.2578125, -0.1708984375, -0.3251953125, -0.5205078125, 0.55615234375, 0.3408203125, 0.02294921875, 0.3076171875, -0.2763671875, -1.349609375, -0.156982421875, 0.0428466796875, -0.63330078125, 0.3662109375, 0.60888671875, -0.075927734375, 0.196533203125, -0.1463623046875, 0.0762939453125, 0.11279296875, 0.001434326171875, -0.4970703125, 0.30615234375, -0.1357421875, -0.459716796875, 0.20751953125, -0.1341552734375, -0.2744140625, -0.174560546875, -0.2137451171875, -0.274658203125, -0.385009765625, 0.0302734375, -0.0133056640625, -0.365234375, -0.5048828125, -0.141357421875, -0.46044921875, -0.1185302734375, 0.22314453125, -0.1199951171875, -0.350341796875, -0.240234375, -0.17724609375, -0.1923828125, 0.179931640625, 0.4921875, 0.12548828125, 0.216552734375, -0.70458984375, -0.3046875, -0.54638671875, -0.1943359375, 0.0074462890625, 0.150390625, 0.1383056640625, 0.0830078125, 0.4130859375, 0.027801513671875, 0.16650390625, -0.0050048828125, -0.1715087890625, 0.416015625, 0.2001953125, 0.572265625, -0.205810546875, 0.04052734375, -0.77294921875, -0.0968017578125, 0.0355224609375, 0.135009765625, -0.2142333984375, -0.0977783203125, -0.5263671875, 0.330322265625, -1.447265625, 0.10382080078125, -0.10357666015625, 1.2451171875, -0.1910400390625, 0.3896484375, 0.420166015625, 0.381103515625, -0.205078125, -0.07110595703125, 0.27734375, -0.0693359375, 0.394287109375, 0.5810546875, -0.5927734375, -0.27099609375, -0.141845703125, -0.193359375, -0.04888916015625, 0.148193359375, 0.48193359375, 0.0667724609375, -0.5458984375, 0.200439453125, -0.188232421875, -0.0677490234375, -0.7763671875, 0.91796875, 0.0029296875, 0.04541015625, -0.0386962890625, -0.6337890625, 0.429931640625, 0.25244140625, 0.5234375, -0.050048828125, -0.638671875, 0.6650390625, 0.25732421875, -0.10888671875, 1.0869140625, 0.043701171875, -0.081787109375, -0.77490234375, -0.2030029296875, -0.307373046875, -0.0047607421875, 0.1904296875, 0.0372314453125, 0.41357421875, -0.196044921875, 0.12939453125, 0.0032958984375, -0.4755859375, -0.3232421875, 0.273193359375, -0.4296875, -0.306396484375, 0.10125732421875, -0.3916015625, -0.141357421875, -0.097900390625, -0.02496337890625, 0.9423828125, 0.00177001953125, -0.30615234375, -0.10546875, 0.16552734375, 0.00994873046875, -0.0340576171875, 0.34912109375, -0.1201171875, 0.244384765625, 0.5810546875, -0.46728515625, 0.215576171875, -0.29345703125, -0.802734375, 0.15625, 0.2454833984375, 0.10546875, 0.0423583984375, 0.0654296875, 0.1396484375, 0.06524658203125, -0.22265625, 0.443115234375, 0.708984375]}, {"frame_number": 5, "vector": [-0.1905517578125, 0.27978515625, 0.015625, -0.2110595703125, -0.6416015625, 0.328857421875, 0.02655029296875, -0.48681640625, -0.080078125, -0.0139312744140625, -0.14990234375, 0.05859375, 0.27685546875, 0.18017578125, -0.029296875, 0.322509765625, 0.1158447265625, -0.134033203125, -0.092041015625, -0.48828125, -0.56201171875, -6.0859375, -0.38134765625, -0.671875, -0.007415771484375, -0.07415771484375, -0.22265625, -0.80859375, 0.513671875, -0.19775390625, -0.95068359375, -0.64990234375, -0.7646484375, 0.03631591796875, -0.069580078125, -0.031097412109375, 0.30322265625, -0.6181640625, -0.08642578125, -0.436279296875, -0.07073974609375, 0.2147216796875, 0.2196044921875, 0.0452880859375, 0.97607421875, -0.35498046875, -0.356689453125, -0.23291015625, -0.580078125, -0.181640625, 0.005641937255859375, -0.2626953125, 0.12646484375, 0.5126953125, 0.244384765625, 0.0113067626953125, 0.400634765625, -0.274169921875, 0.28662109375, -0.010986328125, -0.115478515625, 0.2388916015625, 0.305419921875, 0.184326171875, -0.261474609375, -0.29833984375, -0.474853515625, -0.037841796875, 0.1973876953125, 0.07208251953125, 0.970703125, -0.08544921875, -0.0166015625, -0.0455322265625, -0.448974609375, 0.8857421875, -0.34716796875, -0.45263671875, 0.202392578125, -0.6357421875, 1.1318359375, 0.313232421875, 0.60400390625, 0.5859375, -0.265869140625, 0.05615234375, 0.1832275390625, 0.03204345703125, 0.1591796875, -0.58642578125, -0.1431884765625, -0.133056640625, 0.0599365234375, -0.07421875, 0.23681640625, 0.3232421875, 0.124755859375, 0.57373046875, 0.08380126953125, -0.70166015625, 0.7021484375, 0.30322265625, -0.3564453125, -0.669921875, -0.03814697265625, -0.09033203125, 0.1043701171875, -0.174072265625, -0.1954345703125, -0.04052734375, -0.1123046875, 0.1708984375, 0.0538330078125, -0.414794921875, 0.1435546875, -0.6142578125, -0.175048828125, -0.1483154296875, 0.308349609375, 0.5068359375, 0.040771484375, -0.253662109375, -0.05126953125, -0.6591796875, -0.0902099609375, 0.1710205078125, -0.16943359375, 0.382080078125, 0.1632080078125, 0.34228515625, -0.0107421875, -0.422607421875, -0.0390625, 0.03619384765625, 0.51171875, -0.051025390625, 0.0086669921875, 0.74853515625, 0.578125, 0.026611328125, 0.00018310546875, 0.301513671875, 0.15966796875, -0.097900390625, -0.2783203125, 0.12255859375, 0.01739501953125, -0.3623046875, 0.1900634765625, 0.98779296875, -0.162841796875, -1.552734375, 0.083251953125, -0.42333984375, 0.449462890625, -0.194091796875, -0.27197265625, 0.96435546875, 0.29443359375, -0.418212890625, -0.2352294921875, 0.2578125, 0.486083984375, 0.29736328125, -0.3173828125, -0.26171875, -0.001708984375, 0.6884765625, 0.042724609375, 0.0545654296875, 0.12451171875, 0.363525390625, -0.1436767578125, 0.73779296875, -0.283203125, 1.029296875, -1.16015625, 0.10302734375, -0.0030517578125, -0.062255859375, 0.042724609375, 0.213623046875, -0.1483154296875, 0.03851318359375, 0.0335693359375, -0.00341796875, -0.025634765625, 0.20556640625, -0.0860595703125, 0.141357421875, 0.1937255859375, 0.18994140625, 0.44921875, -0.40576171875, 0.5205078125, -0.26171875, 0.50732421875, -0.230712890625, -0.380859375, 1.1748046875, 0.501953125, -0.04638671875, -0.270751953125, 0.049072265625, 0.0294342041015625, 0.09710693359375, 0.6748046875, 0.047088623046875, -0.69189453125, -0.1859130859375, 0.170654296875, -0.12109375, 0.251708984375, -0.7939453125, -0.49951171875, -0.54833984375, 0.50390625, 1.1220703125, 0.0225982666015625, -0.44189453125, -0.435791015625, -0.564453125, -0.196044921875, -0.6083984375, -0.169189453125, 0.156982421875, -0.0870361328125, -0.236328125, 0.74853515625, 0.03192138671875, -0.1595458984375, 0.2783203125, -0.0177001953125, 0.3125, -0.32177734375, -0.583984375, 0.77783203125, -0.1083984375, -0.92529296875, -0.1123046875, 0.70751953125, 0.268798828125, 0.04541015625, 0.0791015625, -0.2176513671875, 1.03515625, -0.18212890625, 0.00830078125, -0.41357421875, -0.1845703125, 0.0609130859375, -0.01849365234375, 0.2890625, -0.03173828125, -0.67529296875, -0.1015625, -0.1336669921875, -0.03515625, 0.48046875, 0.5185546875, 0.00238037109375, -0.2103271484375, -0.08203125, 0.010009765625, -0.235107421875, 0.045166015625, -0.8466796875, -0.2607421875, 0.58203125, -0.427490234375, 0.109619140625, 0.38232421875, 0.336669921875, 0.2646484375, -0.1812744140625, -0.68017578125, -0.0943603515625, 0.123046875, -0.3466796875, -0.465576171875, -0.312255859375, 0.66259765625, -0.414794921875, 0.11572265625, 0.42724609375, 0.35400390625, -0.1649169921875, 0.5478515625, -0.55908203125, 1.146484375, -0.023101806640625, 0.055999755859375, 0.0546875, -0.33544921875, 0.082763671875, 0.178955078125, 0.1357421875, 0.380126953125, 0.626953125, -0.254638671875, -0.01297760009765625, 0.82275390625, 0.00608062744140625, -0.083984375, -0.312255859375, 0.3564453125, 0.298828125, -0.68896484375, 0.1396484375, -0.084228515625, 0.334716796875, -0.3798828125, -0.01611328125, -0.29248046875, -0.3173828125, 0.4296875, -0.07073974609375, 0.085693359375, -0.256591796875, 0.177978515625, -0.92626953125, 0.135986328125, 0.08282470703125, -0.484619140625, -0.83447265625, 0.60693359375, -0.392578125, 0.28564453125, -0.60400390625, -0.02789306640625, -0.5615234375, -0.08349609375, -0.1649169921875, -0.46923828125, 0.177734375, 0.2415771484375, 0.09442138671875, -0.677734375, 0.059478759765625, 0.0030517578125, -0.6845703125, 0.068359375, -0.267333984375, -0.71728515625, -0.56640625, 0.232666015625, -0.47705078125, 0.38134765625, 0.26220703125, 0.05322265625, -0.35986328125, -0.288330078125, 0.324462890625, 0.103515625, 0.4189453125, 0.332763671875, 0.255859375, -0.121826171875, -0.318359375, -0.1793212890625, -0.213623046875, 0.080810546875, 0.5009765625, 0.3154296875, -0.6064453125, -0.496337890625, -0.03271484375, 0.25439453125, 0.019195556640625, -0.05706787109375, -0.04559326171875, 0.2493896484375, -0.217529296875, 0.3310546875, 0.32763671875, -0.1063232421875, 0.40771484375, -0.281005859375, 0.1572265625, -0.560546875, -0.095703125, -0.348876953125, 0.1234130859375, 0.72314453125, 0.208740234375, -0.059326171875, -0.04559326171875, -0.101806640625, -0.25390625, 0.5634765625, -0.638671875, -0.251953125, 0.267822265625, 0.29296875, 1.0546875, 0.72900390625, -0.19091796875, -0.64306640625, -0.01318359375, -0.408203125, 0.2236328125, 0.1961669921875, -0.1090087890625, -0.45458984375, -0.17236328125, 0.03076171875, -0.372314453125, 0.0531005859375, 0.39013671875, -0.014892578125, 0.251953125, -0.8203125, -0.71875, -0.33203125, -0.423828125, 0.548828125, 0.732421875, -0.28857421875, 1.015625, -0.196533203125, 0.11456298828125, -0.0650634765625, -0.2705078125, -0.0916748046875, -0.0252685546875, -0.5732421875, -1.4267578125, -0.11932373046875, 0.1212158203125, 0.19189453125, 0.05810546875, -6.1015625, -0.37451171875, -0.08740234375, -0.70654296875, -0.2052001953125, -0.072509765625, -1.236328125, -0.6181640625, -0.259765625, -0.04437255859375, 0.6005859375, 0.67138671875, -0.27099609375, -0.0233154296875, 0.28759765625, 0.244140625, 0.0015869140625, -0.1142578125, 0.18505859375, -0.001953125, 0.1695556640625, -0.08062744140625, 0.0770263671875, 0.462158203125, -0.26416015625, -0.19091796875, 0.4140625, -0.6357421875, -0.62451171875, -0.07568359375, 0.151611328125, -0.292236328125, -0.32568359375, -0.51611328125, -0.0419921875, 0.5068359375, 0.77685546875, -0.333984375, -0.388671875, 0.2352294921875, -1.037109375, 0.287841796875, -0.0526123046875, -0.06842041015625, 0.26171875, 0.363525390625, -0.08306884765625, -0.81982421875, -0.8291015625, 0.134765625, 0.037353515625, -1.1171875, -0.010009765625, 0.390869140625, -0.0557861328125, 0.226318359375, -0.1094970703125, 0.1357421875, 0.175048828125, 0.249267578125, 0.6044921875, 0.31201171875, 0.364501953125, -0.45849609375, 0.05908203125, 0.1263427734375, 0.060546875, 0.099365234375, 0.1134033203125, -0.08056640625, 0.063720703125, 0.6455078125, -0.1708984375, -0.26123046875, -0.09521484375, -0.0653076171875, 0.1715087890625, -0.16162109375, 0.092041015625, 0.196044921875, 0.0457763671875, 0.158203125, 0.382568359375, -0.19580078125, -0.126708984375, -0.078125, -0.345458984375, -0.61181640625, -0.127685546875, 0.7705078125, 0.167236328125, -0.169189453125, 0.045440673828125, 0.2158203125, 1.234375, 0.1455078125, 0.171630859375, 0.133544921875, 0.083740234375, -0.3876953125, -0.48095703125, -0.010986328125, -0.044921875, 0.38525390625, 0.348388671875, 0.059326171875, -0.03558349609375, 0.4208984375, -0.0482177734375, 0.7353515625, 0.1396484375, 0.390380859375, 0.479736328125, -0.2386474609375, 0.49853515625, -1.486328125, -0.37548828125, -0.230224609375, 0.1402587890625, 0.00634765625, -0.35546875, 0.76513671875, 0.2802734375, -0.095458984375, 0.10791015625, 0.3603515625, -0.0509033203125, -0.244384765625, -0.90771484375, -0.464111328125, -0.3056640625, -0.455322265625, -0.94140625, 0.24609375, 0.26318359375, 0.6298828125, -0.4033203125, -0.09503173828125, 0.08990478515625, -0.0838623046875, 0.3740234375, -0.251953125, 0.135498046875, 0.212890625, 0.342041015625, -0.404052734375, 0.200439453125, -0.322998046875, -0.493408203125, -0.2919921875, 0.09814453125, -0.15234375, -0.1102294921875, -0.00689697265625, -0.053466796875, 0.0009765625, 0.4853515625, -0.3291015625, 0.205322265625, 0.38427734375, -0.01220703125, -0.169189453125, 0.012969970703125, 0.121826171875, 0.0810546875, -0.0711669921875, -0.182861328125, -1.4833984375, -0.454345703125, 0.1973876953125, -0.06884765625, -0.336669921875, -0.35498046875, -0.192138671875, -0.1480712890625, -0.04205322265625, -0.29150390625, -0.22900390625, -0.8984375, 0.205810546875, 0.389892578125, -0.23291015625, -0.177978515625, -0.1107177734375, -0.5556640625, 0.27490234375, 0.219970703125, 0.005859375, -0.278076171875, 0.02593994140625, -0.06005859375, 0.04461669921875, 0.0738525390625, -0.29248046875, 0.01068115234375, 0.03759765625, 0.38818359375, 0.278076171875, -0.01190185546875, 0.17919921875, 0.019073486328125, -0.36767578125, 0.68505859375, 0.04180908203125, -0.126708984375, -0.330322265625, 1.291015625, 0.092529296875, -0.263916015625, 0.0311279296875, -0.46533203125, 0.362548828125, -0.22802734375, -0.19921875, 0.322265625, 0.1297607421875, 0.0052490234375, 0.259765625, 0.1021728515625, 0.2333984375, -0.33056640625, 0.36376953125, 0.20166015625, -0.6240234375, 0.13720703125, 0.28857421875, -0.1285400390625, -0.279296875, -0.260009765625, -0.2646484375, -0.84912109375, 0.00390625, -0.06610107421875, -0.08416748046875, -0.213623046875, 0.141845703125, -0.0697021484375, 0.4970703125, -0.427734375, -1.38671875, -0.1124267578125, -0.0677490234375, 0.175048828125, 0.137451171875, -0.0804443359375, 0.29345703125, 0.1336669921875, -0.1689453125, 0.1959228515625, -0.014404296875, 0.1812744140625, 0.4072265625, 0.035400390625, 0.151611328125, -0.22412109375, 0.297607421875, 0.221435546875, -0.367919921875, -0.634765625, -0.22021484375, -0.464111328125, 0.04931640625, 0.271484375, -0.282470703125, 0.79541015625, -0.05126953125, 0.0545654296875, -0.8623046875, 0.0523681640625, 0.012451171875, 0.2313232421875, -0.57177734375, 0.2861328125, -0.52490234375, 0.427734375, -0.03521728515625, 0.2325439453125, -0.098876953125, -0.4736328125, -0.6240234375, 0.06402587890625, -0.427001953125, 0.455078125, -0.326171875, 0.96728515625, -0.291259765625, 0.0679931640625, -0.0794677734375, -0.0089111328125, -0.466796875, 0.06591796875, 1.3876953125, -0.01556396484375, 0.28271484375, -0.1087646484375, 0.38818359375, -0.98291015625, -0.209228515625, 0.4775390625, 0.0509033203125, -0.257080078125, 0.517578125, 0.1630859375, 0.0859375, 0.146240234375, 0.03302001953125, 0.165771484375, -0.1484375, 0.44482421875, -0.55029296875, 0.013671875, -0.609375, 0.0889892578125, 0.078857421875, -0.41357421875, 0.02587890625, -0.05218505859375, -0.1710205078125, 0.14306640625, -0.8671875, -0.273193359375, 0.324951171875, -0.132568359375, 0.233642578125, 0.01995849609375, -0.0921630859375, -0.814453125, 0.0196533203125, 0.28662109375, 0.0333251953125, 0.11993408203125, -0.299072265625, -0.51953125, 0.578125, -0.42578125, 0.00921630859375, -0.56298828125, 0.0008544921875, -0.343994140625, 0.1536865234375, 0.0458984375, 0.0687255859375, 0.071044921875, 0.170166015625, -2.35546875, 0.358154296875, -0.173095703125, 0.2802734375, 0.61669921875, -0.0426025390625, -0.31787109375, -0.22705078125, 0.20361328125, -0.1236572265625, 0.254638671875, -0.069091796875, 0.044189453125, 0.439453125, 0.43603515625, 0.319091796875, 0.0318603515625, 0.2490234375, 0.6123046875, -0.259033203125, 0.4365234375, 0.0731201171875, 0.110107421875, -0.28857421875, 0.046875, 0.30126953125, 0.248046875, -0.701171875, 0.495849609375, 0.0589599609375, -0.1461181640625, -0.0206298828125, -0.5048828125, -0.0303955078125, -0.0467529296875, 0.36328125, -0.2420654296875, -0.014190673828125, -0.1904296875, -0.14453125, 0.380615234375, -0.06707763671875, -0.2587890625, -0.65380859375, -0.423095703125, 0.326904296875, 0.0892333984375, -0.187744140625, 0.343505859375, -0.1304931640625, 0.3271484375, -0.446533203125, -0.31298828125, 0.1805419921875, -0.261962890625, -0.4140625, -0.65478515625, -0.654296875, -0.02001953125, 0.802734375, 0.6787109375, -0.2724609375, -0.0057373046875, 0.66259765625, -0.66552734375, 0.19287109375, -0.72998046875, -0.68603515625, -0.464111328125, 0.17529296875, 0.0614013671875, -0.399658203125, -0.33642578125, 0.1405029296875, -0.13623046875, 0.1865234375, 0.484619140625, 0.1610107421875, 0.150146484375, 0.0110626220703125, -0.50927734375, 0.056396484375, -0.11083984375, -0.0906982421875, 0.79443359375, 0.093505859375, -0.287841796875, 0.4951171875, -0.6328125, 0.45458984375, -0.37890625, 0.5380859375, -0.1634521484375, 0.09326171875, 0.1312255859375, -1.021484375, -0.15087890625, 0.03759765625, 0.197265625, -0.29931640625, -0.459716796875, 0.0296630859375, 1.107421875, 0.409912109375, -0.71533203125, -0.255859375, 0.77880859375, -0.1611328125, -0.21484375, -0.10113525390625, -0.3955078125, 0.204833984375, -0.2098388671875, -0.45458984375, 0.411865234375, 0.4287109375, -0.33447265625, -0.505859375, 0.0806884765625, 0.0621337890625, -0.10430908203125, 0.135498046875, -0.296875, -0.06817626953125, 0.24560546875, 0.1763916015625, 0.414306640625, 0.389892578125, 0.255859375, -0.1397705078125, 0.1708984375, -0.00750732421875, 0.3662109375, 0.373046875, 0.51220703125, -0.278564453125, -0.11468505859375, 0.130859375, -0.29296875, -0.02392578125, -0.193115234375, 0.3515625, 0.422607421875, -0.28125, 0.426513671875, -2.12890625, 0.48388671875, 0.45556640625, 0.109619140625, -0.53173828125, -0.18212890625, 0.0941162109375, -0.17041015625, 0.076416015625, 0.15771484375, 0.40625, -0.160888671875, 0.7685546875, -0.379638671875, 0.079833984375, 0.6943359375, 0.27587890625, 0.27490234375, -0.20556640625, -0.3720703125, 0.17236328125, 0.4833984375, 0.155517578125, -0.043609619140625, -0.251708984375, 0.2529296875, 0.1514892578125, 0.99853515625, 0.2313232421875, 0.3271484375, -0.221923828125, 0.320556640625, 0.113037109375, 0.0601806640625, 0.69873046875, -0.75634765625, -0.279052734375, 0.53076171875, -0.6015625, 0.0648193359375, 0.7490234375, -0.462890625, -0.26025390625, -0.222412109375, -0.39404296875, -0.129150390625, 0.0732421875, -0.9130859375, -0.078125, 0.10736083984375, 0.0330810546875, 0.6181640625, -0.31591796875, 0.479736328125, -0.30078125, -0.2054443359375, -0.155029296875, -0.09716796875, 0.05078125, -0.14404296875, 0.1683349609375, 0.05999755859375, 0.134765625, 0.34814453125, -0.224609375, -0.1591796875, -0.36083984375, 0.04144287109375, 0.47265625, 1.2724609375, -0.1566162109375, -0.353271484375, -0.517578125, 0.568359375, 0.33203125, 0.0164794921875, 0.276123046875, -0.28369140625, -1.349609375, -0.143310546875, 0.0517578125, -0.640625, 0.37548828125, 0.60302734375, -0.07843017578125, 0.17041015625, -0.12890625, 0.091064453125, 0.11669921875, 0.00555419921875, -0.5185546875, 0.3056640625, -0.12890625, -0.4287109375, 0.2393798828125, -0.111328125, -0.292724609375, -0.18994140625, -0.2227783203125, -0.252197265625, -0.455322265625, 0.035888671875, -0.0391845703125, -0.385009765625, -0.515625, -0.150634765625, -0.45654296875, -0.131103515625, 0.2447509765625, -0.1177978515625, -0.348876953125, -0.2197265625, -0.18017578125, -0.1895751953125, 0.1552734375, 0.48828125, 0.15771484375, 0.218505859375, -0.70654296875, -0.291015625, -0.55126953125, -0.1817626953125, 0.0185546875, 0.17626953125, 0.13134765625, 0.09515380859375, 0.3994140625, 0.019775390625, 0.1541748046875, -0.02337646484375, -0.17578125, 0.3896484375, 0.228515625, 0.56689453125, -0.187744140625, 0.03271484375, -0.77587890625, -0.07611083984375, 0.0172119140625, 0.1363525390625, -0.2088623046875, -0.09173583984375, -0.5322265625, 0.313720703125, -1.41015625, 0.113037109375, -0.08294677734375, 1.283203125, -0.1910400390625, 0.411865234375, 0.41455078125, 0.384521484375, -0.1942138671875, -0.0595703125, 0.27587890625, -0.092529296875, 0.380126953125, 0.578125, -0.591796875, -0.23974609375, -0.1607666015625, -0.2037353515625, -0.0693359375, 0.166748046875, 0.48583984375, 0.0882568359375, -0.55126953125, 0.216552734375, -0.17578125, -0.0653076171875, -0.76171875, 0.8857421875, 0.008056640625, 0.0140380859375, -0.04052734375, -0.646484375, 0.42919921875, 0.2371826171875, 0.5224609375, -0.0517578125, -0.6220703125, 0.615234375, 0.2734375, -0.142578125, 1.1083984375, 0.0272216796875, -0.0792236328125, -0.7568359375, -0.19580078125, -0.276123046875, -0.01416015625, 0.171875, 0.047119140625, 0.43017578125, -0.18701171875, 0.1312255859375, 0.003173828125, -0.484375, -0.318603515625, 0.223876953125, -0.4091796875, -0.27587890625, 0.1258544921875, -0.393798828125, -0.1534423828125, -0.07147216796875, 0.0218505859375, 0.9580078125, 0.01617431640625, -0.33251953125, -0.1170654296875, 0.1717529296875, 0.032196044921875, -0.0299072265625, 0.364013671875, -0.11029052734375, 0.2225341796875, 0.56103515625, -0.5048828125, 0.26318359375, -0.306884765625, -0.8134765625, 0.1453857421875, 0.257080078125, 0.083984375, 0.050537109375, 0.0833740234375, 0.135009765625, 0.06866455078125, -0.2216796875, 0.429443359375, 0.72607421875]}, {"frame_number": 6, "vector": [-0.170654296875, 0.317138671875, -0.0068359375, -0.217529296875, -0.62353515625, 0.3359375, 0.01104736328125, -0.455322265625, -0.12109375, -0.0233154296875, -0.148193359375, 0.022216796875, 0.32861328125, 0.195556640625, -0.01953125, 0.33935546875, 0.133544921875, -0.1259765625, -0.078369140625, -0.466796875, -0.56298828125, -6.109375, -0.3291015625, -0.68505859375, 0.007598876953125, -0.065185546875, -0.14990234375, -0.783203125, 0.49853515625, -0.2032470703125, -0.9599609375, -0.63427734375, -0.7900390625, 0.06427001953125, -0.0697021484375, 0.02301025390625, 0.297607421875, -0.6357421875, -0.09246826171875, -0.4482421875, -0.09088134765625, 0.1932373046875, 0.2119140625, 0.06640625, 0.89501953125, -0.351806640625, -0.374755859375, -0.22314453125, -0.560546875, -0.17822265625, 0.00670623779296875, -0.246337890625, 0.0882568359375, 0.50390625, 0.253173828125, 0.06182861328125, 0.3779296875, -0.289794921875, 0.2685546875, -0.03680419921875, -0.12451171875, 0.212890625, 0.32470703125, 0.185302734375, -0.2900390625, -0.32666015625, -0.488037109375, -0.052978515625, 0.18212890625, 0.09320068359375, 0.978515625, -0.086181640625, -0.021484375, -0.0166015625, -0.4384765625, 0.84912109375, -0.370361328125, -0.46435546875, 0.18896484375, -0.6533203125, 1.126953125, 0.333984375, 0.5634765625, 0.55322265625, -0.306396484375, 0.0657958984375, 0.17724609375, 0.02685546875, 0.1416015625, -0.609375, -0.16259765625, -0.10723876953125, 0.0615234375, -0.138671875, 0.2626953125, 0.34716796875, 0.08642578125, 0.55078125, 0.06689453125, -0.6796875, 0.7412109375, 0.27734375, -0.37890625, -0.6669921875, -0.0008544921875, -0.1337890625, 0.04376220703125, -0.156005859375, -0.194091796875, -0.059326171875, -0.0888671875, 0.166259765625, 0.09100341796875, -0.448486328125, 0.1502685546875, -0.5810546875, -0.19921875, -0.1346435546875, 0.31591796875, 0.5517578125, 0.0364990234375, -0.2308349609375, -0.0321044921875, -0.7060546875, -0.1158447265625, 0.171875, -0.166259765625, 0.386474609375, 0.1871337890625, 0.39208984375, -0.008544921875, -0.406982421875, -0.0201416015625, 0.07061767578125, 0.5146484375, -0.042724609375, 0.04925537109375, 0.779296875, 0.59326171875, 0.01904296875, 0.05078125, 0.285888671875, 0.133056640625, -0.086669921875, -0.23779296875, 0.0850830078125, -0.017730712890625, -0.350830078125, 0.1978759765625, 0.9697265625, -0.15380859375, -1.5224609375, 0.0601806640625, -0.39404296875, 0.434814453125, -0.2125244140625, -0.2744140625, 0.9443359375, 0.3017578125, -0.41357421875, -0.258544921875, 0.281494140625, 0.48291015625, 0.317138671875, -0.34228515625, -0.277587890625, -0.005859375, 0.7216796875, 0.0419921875, 0.1103515625, 0.1143798828125, 0.38818359375, -0.166748046875, 0.6982421875, -0.284423828125, 1.0693359375, -1.130859375, 0.0933837890625, 0.001220703125, -0.07666015625, 0.0224609375, 0.21337890625, -0.1483154296875, 0.04583740234375, 0.0364990234375, -0.00311279296875, -0.017822265625, 0.2056884765625, -0.0780029296875, 0.136962890625, 0.208740234375, 0.1793212890625, 0.4580078125, -0.357421875, 0.54931640625, -0.2293701171875, 0.49658203125, -0.263916015625, -0.39599609375, 1.19921875, 0.49560546875, -0.05078125, -0.2353515625, 0.0972900390625, -0.02154541015625, 0.09649658203125, 0.65869140625, 0.009429931640625, -0.67626953125, -0.1552734375, 0.172607421875, -0.097412109375, 0.28271484375, -0.82763671875, -0.513671875, -0.5302734375, 0.498291015625, 1.083984375, -0.00884246826171875, -0.380859375, -0.4248046875, -0.56591796875, -0.19775390625, -0.6005859375, -0.16845703125, 0.130615234375, -0.1253662109375, -0.23583984375, 0.7802734375, 0.025634765625, -0.1070556640625, 0.273681640625, 0.0196533203125, 0.294189453125, -0.322265625, -0.54541015625, 0.7978515625, -0.073486328125, -0.9296875, -0.119140625, 0.6669921875, 0.2364501953125, 0.0745849609375, 0.1005859375, -0.1781005859375, 0.99609375, -0.1640625, 0.0162353515625, -0.4189453125, -0.1337890625, 0.0443115234375, -0.04046630859375, 0.30859375, -0.010498046875, -0.70263671875, -0.0859375, -0.17919921875, -0.014892578125, 0.461669921875, 0.485595703125, -0.0042724609375, -0.1973876953125, -0.06982421875, -0.0068359375, -0.2509765625, 0.03277587890625, -0.837890625, -0.25390625, 0.59765625, -0.4677734375, 0.06097412109375, 0.42236328125, 0.33740234375, 0.238525390625, -0.160888671875, -0.68603515625, -0.080078125, 0.12890625, -0.36083984375, -0.42919921875, -0.3076171875, 0.6630859375, -0.38916015625, 0.120849609375, 0.423828125, 0.338134765625, -0.1685791015625, 0.5546875, -0.5341796875, 1.1259765625, -0.021209716796875, 0.06036376953125, 0.0201416015625, -0.36279296875, 0.0921630859375, 0.21142578125, 0.147216796875, 0.34423828125, 0.6162109375, -0.259521484375, -0.0012969970703125, 0.787109375, -0.0105438232421875, -0.052490234375, -0.295166015625, 0.357177734375, 0.31884765625, -0.71533203125, 0.129638671875, -0.0501708984375, 0.3603515625, -0.359375, -0.03076171875, -0.294921875, -0.345947265625, 0.4169921875, -0.058013916015625, 0.0413818359375, -0.25390625, 0.178955078125, -0.9013671875, 0.0928955078125, 0.110595703125, -0.4775390625, -0.84912109375, 0.65234375, -0.400390625, 0.31884765625, -0.60791015625, -0.01702880859375, -0.5400390625, -0.0445556640625, -0.166259765625, -0.50732421875, 0.183837890625, 0.2205810546875, 0.08367919921875, -0.6162109375, 0.0355224609375, 0.0081787109375, -0.69384765625, 0.0423583984375, -0.272216796875, -0.67626953125, -0.54052734375, 0.2099609375, -0.493408203125, 0.40869140625, 0.2275390625, 0.067626953125, -0.320556640625, -0.250244140625, 0.280517578125, 0.14990234375, 0.4443359375, 0.345703125, 0.25537109375, -0.1409912109375, -0.33154296875, -0.1741943359375, -0.1732177734375, 0.05859375, 0.49072265625, 0.3193359375, -0.60546875, -0.471923828125, -0.03466796875, 0.2318115234375, 0.057586669921875, -0.06085205078125, -0.0631103515625, 0.25390625, -0.21240234375, 0.33642578125, 0.314453125, -0.1019287109375, 0.38037109375, -0.2340087890625, 0.1630859375, -0.5419921875, -0.065185546875, -0.404052734375, 0.12158203125, 0.6953125, 0.21630859375, -0.0439453125, -0.0262451171875, -0.111328125, -0.27099609375, 0.544921875, -0.67578125, -0.234130859375, 0.33349609375, 0.30517578125, 1.033203125, 0.6806640625, -0.159912109375, -0.646484375, -0.0489501953125, -0.45166015625, 0.24462890625, 0.19970703125, -0.112060546875, -0.43798828125, -0.2108154296875, 0.0115966796875, -0.362060546875, 0.0712890625, 0.372802734375, -0.015869140625, 0.202880859375, -0.84521484375, -0.71630859375, -0.3359375, -0.396728515625, 0.5322265625, 0.720703125, -0.2890625, 1.04296875, -0.1993408203125, 0.11700439453125, -0.0653076171875, -0.24072265625, -0.106201171875, -0.0379638671875, -0.5458984375, -1.3974609375, -0.12420654296875, 0.0865478515625, 0.197265625, 0.036224365234375, -6.11328125, -0.4150390625, -0.080078125, -0.7587890625, -0.197509765625, -0.1019287109375, -1.2109375, -0.6513671875, -0.2391357421875, 0.009521484375, 0.626953125, 0.66943359375, -0.251708984375, -0.04180908203125, 0.211181640625, 0.23046875, 0.000244140625, -0.06201171875, 0.1845703125, 0.00244140625, 0.1824951171875, -0.14453125, 0.068115234375, 0.47119140625, -0.279296875, -0.1590576171875, 0.42529296875, -0.634765625, -0.60888671875, -0.07861328125, 0.12451171875, -0.26953125, -0.338623046875, -0.46826171875, -0.05712890625, 0.51513671875, 0.787109375, -0.328125, -0.400390625, 0.2259521484375, -1.03125, 0.293212890625, -0.03955078125, -0.04248046875, 0.23974609375, 0.323974609375, -0.07794189453125, -0.8115234375, -0.79443359375, 0.146728515625, 0.05615234375, -1.111328125, -0.007568359375, 0.38525390625, -0.00531005859375, 0.22705078125, -0.11572265625, 0.129638671875, 0.1708984375, 0.21484375, 0.6025390625, 0.32861328125, 0.411376953125, -0.5048828125, 0.02734375, 0.0721435546875, 0.0533447265625, 0.093994140625, 0.091796875, -0.067626953125, 0.03533935546875, 0.65869140625, -0.173828125, -0.23828125, -0.10595703125, -0.0716552734375, 0.12158203125, -0.171142578125, 0.0771484375, 0.203125, 0.0494384765625, 0.166259765625, 0.39501953125, -0.19140625, -0.1573486328125, -0.090087890625, -0.341552734375, -0.609375, -0.1488037109375, 0.72265625, 0.18505859375, -0.195556640625, 0.049102783203125, 0.238037109375, 1.2294921875, 0.145263671875, 0.162109375, 0.148193359375, 0.113525390625, -0.355224609375, -0.4775390625, 0.016845703125, -0.031494140625, 0.34814453125, 0.325439453125, 0.046142578125, -0.05462646484375, 0.39599609375, -0.1044921875, 0.7158203125, 0.1357421875, 0.377197265625, 0.5107421875, -0.220703125, 0.482421875, -1.5068359375, -0.35595703125, -0.214599609375, 0.1280517578125, -0.00244140625, -0.3583984375, 0.75048828125, 0.27099609375, -0.1160888671875, 0.07421875, 0.30126953125, -0.0523681640625, -0.25146484375, -0.95166015625, -0.45654296875, -0.30322265625, -0.427001953125, -0.9638671875, 0.251220703125, 0.299560546875, 0.57763671875, -0.36572265625, -0.09521484375, 0.05743408203125, -0.0460205078125, 0.377197265625, -0.220703125, 0.1466064453125, 0.183837890625, 0.342041015625, -0.36474609375, 0.21484375, -0.333984375, -0.5087890625, -0.3154296875, 0.123046875, -0.16259765625, -0.107177734375, -0.02508544921875, -0.10546875, 0.0234375, 0.478515625, -0.349609375, 0.1934814453125, 0.3603515625, 0.024932861328125, -0.18310546875, 0.00225830078125, 0.09771728515625, 0.057373046875, -0.0615234375, -0.205322265625, -1.4833984375, -0.454833984375, 0.16845703125, -0.05450439453125, -0.31689453125, -0.3544921875, -0.2115478515625, -0.1463623046875, -0.0860595703125, -0.316650390625, -0.197998046875, -0.880859375, 0.256103515625, 0.39013671875, -0.2060546875, -0.191162109375, -0.107421875, -0.5908203125, 0.269287109375, 0.2362060546875, -0.0252685546875, -0.267578125, 0.02374267578125, -0.08544921875, 0.051605224609375, 0.07080078125, -0.260009765625, 0.001007080078125, 0.0654296875, 0.390625, 0.273193359375, -0.05810546875, 0.1551513671875, 0.0469970703125, -0.3896484375, 0.69580078125, 0.08203125, -0.1181640625, -0.31298828125, 1.3212890625, 0.081298828125, -0.271240234375, 0.019775390625, -0.48828125, 0.37109375, -0.2283935546875, -0.203369140625, 0.33935546875, 0.1324462890625, 0.0172119140625, 0.285400390625, 0.0985107421875, 0.237060546875, -0.32666015625, 0.3623046875, 0.203125, -0.6328125, 0.128662109375, 0.312255859375, -0.1185302734375, -0.26953125, -0.26318359375, -0.29296875, -0.86767578125, 0.0218505859375, -0.0618896484375, -0.049346923828125, -0.23681640625, 0.10791015625, -0.053466796875, 0.49560546875, -0.453125, -1.384765625, -0.076416015625, -0.150390625, 0.195068359375, 0.19677734375, -0.076416015625, 0.296875, 0.1552734375, -0.173095703125, 0.2177734375, -0.01220703125, 0.172607421875, 0.416259765625, 0.025146484375, 0.1370849609375, -0.197509765625, 0.30712890625, 0.175048828125, -0.36669921875, -0.62353515625, -0.2110595703125, -0.465087890625, 0.0478515625, 0.248291015625, -0.3125, 0.7939453125, -0.05908203125, 0.0869140625, -0.8544921875, 0.008544921875, -0.0098876953125, 0.2330322265625, -0.6123046875, 0.30419921875, -0.50390625, 0.41162109375, -0.033355712890625, 0.2412109375, -0.1038818359375, -0.5029296875, -0.62890625, 0.13134765625, -0.428466796875, 0.441162109375, -0.34423828125, 0.958984375, -0.304931640625, 0.061279296875, -0.11865234375, -0.0172119140625, -0.46484375, 0.067626953125, 1.3671875, -0.006805419921875, 0.2880859375, -0.1229248046875, 0.402099609375, -1.044921875, -0.225830078125, 0.48291015625, 0.01837158203125, -0.2548828125, 0.51904296875, 0.1646728515625, 0.061279296875, 0.1474609375, 0.07916259765625, 0.150390625, -0.1278076171875, 0.45263671875, -0.52734375, 0.01123046875, -0.63330078125, 0.0465087890625, 0.094970703125, -0.42138671875, 0.0380859375, -0.0667724609375, -0.1197509765625, 0.109130859375, -0.89892578125, -0.248046875, 0.313232421875, -0.106689453125, 0.240966796875, 0.018310546875, -0.09423828125, -0.822265625, 0.0234375, 0.28564453125, 0.016357421875, 0.069091796875, -0.250244140625, -0.53076171875, 0.61328125, -0.400146484375, -0.0203857421875, -0.58837890625, 0.00347900390625, -0.36376953125, 0.158935546875, -0.01513671875, 0.0582275390625, 0.07513427734375, 0.169677734375, -2.30078125, 0.370361328125, -0.19775390625, 0.285400390625, 0.63818359375, -0.044677734375, -0.3037109375, -0.218505859375, 0.22216796875, -0.1195068359375, 0.197509765625, -0.0726318359375, 0.0391845703125, 0.45263671875, 0.48046875, 0.31103515625, 0.01123046875, 0.2415771484375, 0.6181640625, -0.26904296875, 0.4130859375, 0.09039306640625, 0.1224365234375, -0.25341796875, 0.079345703125, 0.266845703125, 0.2099609375, -0.74072265625, 0.52392578125, 0.0791015625, -0.18408203125, -0.0076904296875, -0.5048828125, -0.0054931640625, -0.03564453125, 0.35400390625, -0.2763671875, -0.011444091796875, -0.20458984375, -0.1943359375, 0.380615234375, -0.0440673828125, -0.2421875, -0.642578125, -0.41552734375, 0.302978515625, 0.1038818359375, -0.1947021484375, 0.376708984375, -0.12408447265625, 0.34326171875, -0.44287109375, -0.31201171875, 0.190185546875, -0.311279296875, -0.474853515625, -0.6494140625, -0.6181640625, -0.0306396484375, 0.77587890625, 0.6005859375, -0.243408203125, -0.013916015625, 0.7041015625, -0.6708984375, 0.228271484375, -0.658203125, -0.68505859375, -0.488525390625, 0.15185546875, 0.07318115234375, -0.38232421875, -0.322998046875, 0.177734375, -0.12060546875, 0.154296875, 0.4658203125, 0.1180419921875, 0.11328125, -0.01459503173828125, -0.494384765625, 0.053955078125, -0.0596923828125, -0.0828857421875, 0.8125, 0.115966796875, -0.27099609375, 0.52490234375, -0.662109375, 0.468017578125, -0.359375, 0.5654296875, -0.1402587890625, 0.1041259765625, 0.11572265625, -1.001953125, -0.1640625, 0.05810546875, 0.1927490234375, -0.327880859375, -0.51806640625, 0.019775390625, 1.1005859375, 0.42236328125, -0.6708984375, -0.27392578125, 0.763671875, -0.158935546875, -0.219970703125, -0.12127685546875, -0.40673828125, 0.206298828125, -0.1898193359375, -0.45361328125, 0.439208984375, 0.43505859375, -0.32568359375, -0.501953125, 0.10113525390625, 0.07861328125, -0.1318359375, 0.144287109375, -0.241455078125, -0.055145263671875, 0.222900390625, 0.186767578125, 0.41064453125, 0.373291015625, 0.263427734375, -0.134033203125, 0.161376953125, -0.00970458984375, 0.3681640625, 0.356689453125, 0.509765625, -0.268798828125, -0.11572265625, 0.134033203125, -0.332763671875, 0.02001953125, -0.162841796875, 0.38134765625, 0.402587890625, -0.30126953125, 0.45654296875, -2.140625, 0.5087890625, 0.4365234375, 0.10302734375, -0.52880859375, -0.1754150390625, 0.0758056640625, -0.136962890625, 0.10089111328125, 0.24658203125, 0.351806640625, -0.11865234375, 0.77197265625, -0.384765625, 0.0806884765625, 0.69970703125, 0.25341796875, 0.27099609375, -0.17724609375, -0.3603515625, 0.2017822265625, 0.49951171875, 0.1556396484375, -0.0802001953125, -0.188720703125, 0.26123046875, 0.1278076171875, 0.9873046875, 0.28271484375, 0.29248046875, -0.2294921875, 0.297607421875, 0.17724609375, 0.01580810546875, 0.63623046875, -0.74169921875, -0.290771484375, 0.54296875, -0.63720703125, 0.0445556640625, 0.75146484375, -0.4599609375, -0.2626953125, -0.14794921875, -0.40478515625, -0.2197265625, 0.1136474609375, -0.94189453125, -0.08935546875, 0.1431884765625, 0.0228271484375, 0.6279296875, -0.26416015625, 0.423095703125, -0.3193359375, -0.20458984375, -0.1904296875, -0.0888671875, 0.072509765625, -0.09033203125, 0.1192626953125, 0.0565185546875, 0.1485595703125, 0.355712890625, -0.193115234375, -0.1552734375, -0.359130859375, 0.050933837890625, 0.4423828125, 1.224609375, -0.1519775390625, -0.32568359375, -0.595703125, 0.5732421875, 0.330078125, 0.0113525390625, 0.33203125, -0.257080078125, -1.349609375, -0.0994873046875, 0.02880859375, -0.6162109375, 0.378173828125, 0.5947265625, -0.07275390625, 0.18603515625, -0.1475830078125, 0.0941162109375, 0.146484375, 0.00872802734375, -0.48974609375, 0.3115234375, -0.1458740234375, -0.44482421875, 0.179931640625, -0.1148681640625, -0.293701171875, -0.1842041015625, -0.2337646484375, -0.284912109375, -0.4462890625, 0.027587890625, -0.0079345703125, -0.38525390625, -0.56787109375, -0.130126953125, -0.47314453125, -0.1151123046875, 0.214599609375, -0.091796875, -0.341796875, -0.21435546875, -0.17041015625, -0.203125, 0.169921875, 0.50634765625, 0.162841796875, 0.2103271484375, -0.70849609375, -0.310546875, -0.54443359375, -0.2159423828125, 0.0244140625, 0.1871337890625, 0.1363525390625, 0.07781982421875, 0.426025390625, -0.005950927734375, 0.164306640625, -0.0218505859375, -0.173095703125, 0.41845703125, 0.23291015625, 0.568359375, -0.19091796875, 0.0341796875, -0.7890625, -0.1114501953125, 0.052734375, 0.160888671875, -0.2239990234375, -0.08099365234375, -0.5283203125, 0.3134765625, -1.443359375, 0.1290283203125, -0.128173828125, 1.2646484375, -0.169677734375, 0.391357421875, 0.373046875, 0.4453125, -0.184814453125, -0.0513916015625, 0.265869140625, -0.0479736328125, 0.434326171875, 0.56982421875, -0.62109375, -0.2841796875, -0.1439208984375, -0.18701171875, -0.0426025390625, 0.1475830078125, 0.46142578125, 0.08349609375, -0.603515625, 0.18994140625, -0.199462890625, -0.0770263671875, -0.7998046875, 0.88427734375, 0.00390625, -0.00537109375, -0.03125, -0.666015625, 0.449951171875, 0.219482421875, 0.51123046875, -0.051025390625, -0.6220703125, 0.68505859375, 0.2958984375, -0.09747314453125, 1.11328125, 0.06103515625, -0.0784912109375, -0.7890625, -0.2178955078125, -0.302001953125, 0.009033203125, 0.16064453125, 0.035369873046875, 0.4150390625, -0.189453125, 0.1339111328125, -0.0185546875, -0.4658203125, -0.32373046875, 0.234375, -0.42626953125, -0.2841796875, 0.10772705078125, -0.363525390625, -0.12353515625, -0.1119384765625, -0.03887939453125, 0.935546875, -0.036376953125, -0.31298828125, -0.10400390625, 0.17529296875, 0.0511474609375, -0.02001953125, 0.30419921875, -0.10565185546875, 0.255126953125, 0.55322265625, -0.48583984375, 0.2509765625, -0.30419921875, -0.80029296875, 0.14794921875, 0.29443359375, 0.087158203125, -0.00616455078125, 0.078857421875, 0.1484375, 0.0989990234375, -0.225830078125, 0.45361328125, 0.71337890625]}, {"frame_number": 7, "vector": [-0.14453125, 0.308837890625, -0.00146484375, -0.1895751953125, -0.6484375, 0.343505859375, -0.0343017578125, -0.480224609375, -0.1441650390625, -0.0022430419921875, -0.16455078125, 0.02197265625, 0.35888671875, 0.163818359375, -0.0330810546875, 0.315673828125, 0.130615234375, -0.080322265625, -0.08917236328125, -0.45849609375, -0.5361328125, -6.1328125, -0.34521484375, -0.68994140625, -0.00091552734375, -0.042694091796875, -0.17236328125, -0.76953125, 0.4970703125, -0.18896484375, -0.9501953125, -0.6611328125, -0.767578125, 0.0340576171875, -0.048828125, 0.0091552734375, 0.2861328125, -0.65478515625, -0.07476806640625, -0.421630859375, -0.05035400390625, 0.202392578125, 0.246826171875, 0.0595703125, 0.8955078125, -0.3544921875, -0.4052734375, -0.2091064453125, -0.59130859375, -0.185546875, 0.0296478271484375, -0.28857421875, 0.11224365234375, 0.53076171875, 0.241455078125, 0.06268310546875, 0.3974609375, -0.252685546875, 0.2802734375, -0.033447265625, -0.158203125, 0.24755859375, 0.30419921875, 0.185546875, -0.2724609375, -0.34765625, -0.47802734375, -0.0616455078125, 0.18505859375, 0.07293701171875, 0.9619140625, -0.076171875, -0.0390625, -0.0396728515625, -0.440673828125, 0.8583984375, -0.3486328125, -0.4677734375, 0.22119140625, -0.6787109375, 1.1806640625, 0.322021484375, 0.57568359375, 0.5498046875, -0.276611328125, 0.0443115234375, 0.1781005859375, 0.01654052734375, 0.123291015625, -0.59423828125, -0.1680908203125, -0.11224365234375, 0.0638427734375, -0.1474609375, 0.26025390625, 0.355224609375, 0.0758056640625, 0.58447265625, 0.055023193359375, -0.67529296875, 0.73828125, 0.302978515625, -0.3798828125, -0.6416015625, 0.000823974609375, -0.132080078125, 0.03399658203125, -0.131591796875, -0.1614990234375, -0.051513671875, -0.07415771484375, 0.1845703125, 0.098388671875, -0.470947265625, 0.1561279296875, -0.6162109375, -0.248291015625, -0.1591796875, 0.319580078125, 0.53759765625, 0.038330078125, -0.2249755859375, -0.035888671875, -0.72265625, -0.1220703125, 0.18115234375, -0.1441650390625, 0.362548828125, 0.1920166015625, 0.361572265625, -0.038818359375, -0.4013671875, 0.0142822265625, 0.078125, 0.5751953125, -0.046630859375, 0.03363037109375, 0.76416015625, 0.6083984375, 0.010498046875, 0.0662841796875, 0.2666015625, 0.1572265625, -0.089111328125, -0.233642578125, 0.0775146484375, -0.0194091796875, -0.336181640625, 0.19482421875, 1.005859375, -0.1572265625, -1.525390625, 0.0390625, -0.41455078125, 0.415771484375, -0.2198486328125, -0.28271484375, 0.92333984375, 0.26904296875, -0.41748046875, -0.26171875, 0.31884765625, 0.484619140625, 0.315185546875, -0.34033203125, -0.29052734375, -0.000732421875, 0.7109375, 0.013427734375, 0.125244140625, 0.13330078125, 0.414306640625, -0.1795654296875, 0.6630859375, -0.263916015625, 1.037109375, -1.158203125, 0.1102294921875, -0.003662109375, -0.085693359375, 0.041259765625, 0.21484375, -0.1585693359375, 0.0487060546875, 0.057373046875, -0.00604248046875, 0.022216796875, 0.2086181640625, -0.0838623046875, 0.1634521484375, 0.210205078125, 0.1593017578125, 0.44873046875, -0.3564453125, 0.5556640625, -0.207275390625, 0.465087890625, -0.252197265625, -0.388671875, 1.1875, 0.467529296875, -0.05126953125, -0.230224609375, 0.0750732421875, -0.01483154296875, 0.08660888671875, 0.6728515625, -0.0089111328125, -0.6953125, -0.134033203125, 0.198974609375, -0.110595703125, 0.27685546875, -0.82373046875, -0.509765625, -0.525390625, 0.47509765625, 1.06640625, 0.0043792724609375, -0.408935546875, -0.438720703125, -0.56005859375, -0.18408203125, -0.61328125, -0.18017578125, 0.1796875, -0.1175537109375, -0.27734375, 0.806640625, 0.020263671875, -0.12841796875, 0.27978515625, 0.04150390625, 0.2880859375, -0.30810546875, -0.54541015625, 0.80078125, -0.0550537109375, -0.96337890625, -0.11767578125, 0.662109375, 0.234619140625, 0.0340576171875, 0.07861328125, -0.189453125, 1.0048828125, -0.177734375, -0.0037841796875, -0.4150390625, -0.129150390625, 0.0078125, -0.033203125, 0.30029296875, -0.02880859375, -0.6728515625, -0.087158203125, -0.184326171875, -0.02490234375, 0.51171875, 0.490234375, 0.00726318359375, -0.1873779296875, -0.0673828125, 0.01318359375, -0.26708984375, 0.0211181640625, -0.830078125, -0.286865234375, 0.5791015625, -0.451416015625, 0.06298828125, 0.427734375, 0.35400390625, 0.2626953125, -0.165771484375, -0.67626953125, -0.08721923828125, 0.127685546875, -0.348388671875, -0.429443359375, -0.2724609375, 0.642578125, -0.400390625, 0.10693359375, 0.4326171875, 0.341064453125, -0.1513671875, 0.54052734375, -0.53271484375, 1.109375, -0.023956298828125, 0.031280517578125, 0.049072265625, -0.359130859375, 0.08624267578125, 0.208984375, 0.150634765625, 0.364990234375, 0.625, -0.260009765625, -0.01428985595703125, 0.7919921875, 0.00521087646484375, -0.047119140625, -0.2900390625, 0.36572265625, 0.349853515625, -0.70654296875, 0.126220703125, -0.0830078125, 0.364013671875, -0.353759765625, -0.0166015625, -0.323974609375, -0.35498046875, 0.42431640625, -0.07342529296875, 0.00445556640625, -0.27734375, 0.21337890625, -0.90380859375, 0.1112060546875, 0.10107421875, -0.480712890625, -0.87158203125, 0.623046875, -0.444091796875, 0.33740234375, -0.5986328125, -0.0150146484375, -0.525390625, -0.053466796875, -0.1656494140625, -0.491943359375, 0.180908203125, 0.2037353515625, 0.1080322265625, -0.59375, 0.0435791015625, 0.003662109375, -0.69287109375, 0.017822265625, -0.28955078125, -0.720703125, -0.515625, 0.205810546875, -0.45068359375, 0.35791015625, 0.2244873046875, 0.0582275390625, -0.3212890625, -0.299072265625, 0.276123046875, 0.155517578125, 0.435302734375, 0.3681640625, 0.25537109375, -0.14306640625, -0.319091796875, -0.1695556640625, -0.1781005859375, 0.0712890625, 0.49658203125, 0.32080078125, -0.6220703125, -0.463134765625, -0.03662109375, 0.2548828125, 0.0728759765625, -0.05322265625, -0.06201171875, 0.264404296875, -0.18994140625, 0.32421875, 0.309814453125, -0.079833984375, 0.38818359375, -0.22314453125, 0.19580078125, -0.52490234375, -0.080322265625, -0.4150390625, 0.1373291015625, 0.6611328125, 0.232421875, -0.054443359375, -0.041259765625, -0.133056640625, -0.27294921875, 0.560546875, -0.6533203125, -0.229736328125, 0.314453125, 0.28759765625, 1.03515625, 0.67236328125, -0.1689453125, -0.63720703125, -0.0472412109375, -0.47216796875, 0.21728515625, 0.1456298828125, -0.107421875, -0.455078125, -0.173828125, 0.0101318359375, -0.396240234375, 0.075439453125, 0.38525390625, 0.01611328125, 0.2041015625, -0.87060546875, -0.7294921875, -0.32568359375, -0.41162109375, 0.5439453125, 0.7314453125, -0.29345703125, 1.068359375, -0.17822265625, 0.11749267578125, -0.02337646484375, -0.203857421875, -0.116943359375, -0.0501708984375, -0.49609375, -1.3759765625, -0.1317138671875, 0.0740966796875, 0.2042236328125, 0.07464599609375, -6.140625, -0.43017578125, -0.061279296875, -0.73876953125, -0.1864013671875, -0.1114501953125, -1.2001953125, -0.65869140625, -0.2353515625, -0.010498046875, 0.59619140625, 0.66064453125, -0.254150390625, -0.02117919921875, 0.2169189453125, 0.23583984375, 0.008544921875, -0.0400390625, 0.1861572265625, 0.018798828125, 0.1781005859375, -0.12493896484375, 0.06298828125, 0.44775390625, -0.2939453125, -0.1845703125, 0.432373046875, -0.640625, -0.60302734375, -0.079833984375, 0.134033203125, -0.2724609375, -0.347900390625, -0.4443359375, -0.055419921875, 0.53759765625, 0.78515625, -0.34423828125, -0.41162109375, 0.2279052734375, -1.0419921875, 0.311279296875, -0.038818359375, -0.0137939453125, 0.253173828125, 0.31884765625, -0.0858154296875, -0.7939453125, -0.77978515625, 0.1435546875, 0.083251953125, -1.1015625, -0.018798828125, 0.37060546875, -0.0172119140625, 0.232177734375, -0.09765625, 0.1435546875, 0.172119140625, 0.205810546875, 0.59375, 0.334716796875, 0.4150390625, -0.464111328125, 0.018798828125, 0.07000732421875, 0.017333984375, 0.10302734375, 0.0784912109375, -0.0797119140625, 0.05657958984375, 0.6611328125, -0.1707763671875, -0.206787109375, -0.126953125, -0.09307861328125, 0.1297607421875, -0.18212890625, 0.072998046875, 0.203125, 0.04443359375, 0.15234375, 0.38134765625, -0.1884765625, -0.1697998046875, -0.09765625, -0.326171875, -0.57177734375, -0.14111328125, 0.7197265625, 0.18310546875, -0.1953125, 0.056884765625, 0.2237548828125, 1.2529296875, 0.169921875, 0.119873046875, 0.154541015625, 0.089111328125, -0.365234375, -0.476806640625, -0.024658203125, -0.039794921875, 0.36328125, 0.308837890625, 0.05908203125, -0.0216064453125, 0.43896484375, -0.114501953125, 0.7568359375, 0.15185546875, 0.365478515625, 0.53076171875, -0.2054443359375, 0.490966796875, -1.5263671875, -0.39306640625, -0.2176513671875, 0.0859375, -0.0244140625, -0.351806640625, 0.76416015625, 0.288818359375, -0.107177734375, 0.08740234375, 0.29052734375, -0.05859375, -0.239501953125, -0.93017578125, -0.4384765625, -0.28173828125, -0.4638671875, -0.96484375, 0.26904296875, 0.29833984375, 0.5517578125, -0.380859375, -0.07415771484375, 0.04736328125, -0.0523681640625, 0.3486328125, -0.228515625, 0.12158203125, 0.1768798828125, 0.35205078125, -0.367919921875, 0.1904296875, -0.31005859375, -0.491455078125, -0.325439453125, 0.13818359375, -0.16015625, -0.0946044921875, 0.00225830078125, -0.14404296875, 0.0185546875, 0.4736328125, -0.368896484375, 0.20361328125, 0.361083984375, 0.03985595703125, -0.1951904296875, 0.0032958984375, 0.09619140625, 0.0665283203125, -0.0599365234375, -0.2110595703125, -1.474609375, -0.451416015625, 0.163330078125, -0.04901123046875, -0.3310546875, -0.3603515625, -0.240234375, -0.1470947265625, -0.081298828125, -0.29931640625, -0.2315673828125, -0.8779296875, 0.2509765625, 0.36474609375, -0.2127685546875, -0.2236328125, -0.1278076171875, -0.5908203125, 0.248779296875, 0.2449951171875, -0.044403076171875, -0.27099609375, 0.0372314453125, -0.080322265625, 0.0704345703125, 0.06298828125, -0.245849609375, -0.016693115234375, 0.0770263671875, 0.421875, 0.263671875, -0.0673828125, 0.152587890625, 0.059326171875, -0.386474609375, 0.71875, 0.0692138671875, -0.11669921875, -0.3017578125, 1.2958984375, 0.05810546875, -0.2802734375, 0.01708984375, -0.457763671875, 0.35693359375, -0.249755859375, -0.199462890625, 0.347900390625, 0.1380615234375, -0.00689697265625, 0.27294921875, 0.1104736328125, 0.2244873046875, -0.341796875, 0.381103515625, 0.21484375, -0.6484375, 0.14453125, 0.318603515625, -0.1094970703125, -0.26513671875, -0.26318359375, -0.27783203125, -0.8720703125, 0.03271484375, -0.0823974609375, -0.047088623046875, -0.226806640625, 0.10546875, -0.0947265625, 0.486572265625, -0.460205078125, -1.390625, -0.1070556640625, -0.1416015625, 0.190185546875, 0.2052001953125, -0.06689453125, 0.30712890625, 0.16455078125, -0.181640625, 0.2254638671875, -0.0211181640625, 0.1566162109375, 0.438232421875, 0.02490234375, 0.1259765625, -0.1866455078125, 0.30810546875, 0.19775390625, -0.37646484375, -0.62109375, -0.217041015625, -0.47705078125, 0.025390625, 0.2763671875, -0.290771484375, 0.8193359375, -0.0489501953125, 0.09912109375, -0.86083984375, -0.00848388671875, -0.0179443359375, 0.2049560546875, -0.6005859375, 0.312744140625, -0.5234375, 0.43603515625, -0.0318603515625, 0.226318359375, -0.09716796875, -0.52734375, -0.66943359375, 0.12042236328125, -0.40185546875, 0.41748046875, -0.318359375, 0.984375, -0.34814453125, 0.0601806640625, -0.103515625, -0.0064697265625, -0.44873046875, 0.095458984375, 1.365234375, 0.02337646484375, 0.278076171875, -0.121337890625, 0.380859375, -1.0703125, -0.229736328125, 0.484375, 0.0277099609375, -0.23046875, 0.481689453125, 0.173095703125, 0.07177734375, 0.156494140625, 0.078125, 0.1845703125, -0.114501953125, 0.4560546875, -0.52392578125, 0.0302734375, -0.6533203125, 0.0638427734375, 0.1002197265625, -0.4384765625, 0.01556396484375, -0.0970458984375, -0.1328125, 0.1083984375, -0.892578125, -0.2464599609375, 0.25732421875, -0.115966796875, 0.2218017578125, -0.00048828125, -0.102294921875, -0.8076171875, 0.032470703125, 0.2939453125, -0.023681640625, 0.08209228515625, -0.2352294921875, -0.54638671875, 0.5927734375, -0.4267578125, -0.036529541015625, -0.5966796875, -0.01300048828125, -0.35546875, 0.1553955078125, -0.0537109375, 0.0281982421875, 0.0970458984375, 0.1785888671875, -2.28125, 0.364501953125, -0.213623046875, 0.270263671875, 0.6611328125, -0.0682373046875, -0.30419921875, -0.22900390625, 0.20458984375, -0.1328125, 0.224853515625, -0.0343017578125, 0.026611328125, 0.45068359375, 0.477783203125, 0.32080078125, 0.0042724609375, 0.246337890625, 0.638671875, -0.2861328125, 0.400146484375, 0.06134033203125, 0.1322021484375, -0.241455078125, 0.07421875, 0.263671875, 0.23046875, -0.7353515625, 0.5439453125, 0.0863037109375, -0.149658203125, -0.02783203125, -0.52099609375, -0.00341796875, -0.00384521484375, 0.31787109375, -0.2900390625, -0.05340576171875, -0.1953125, -0.181640625, 0.406005859375, -0.05029296875, -0.27001953125, -0.61474609375, -0.41064453125, 0.30126953125, 0.1109619140625, -0.176025390625, 0.377685546875, -0.12396240234375, 0.3291015625, -0.41748046875, -0.278076171875, 0.18505859375, -0.3212890625, -0.470703125, -0.669921875, -0.607421875, -0.0389404296875, 0.787109375, 0.61474609375, -0.24169921875, 0.04498291015625, 0.69287109375, -0.6748046875, 0.246826171875, -0.6845703125, -0.650390625, -0.477294921875, 0.11865234375, 0.0673828125, -0.40087890625, -0.32958984375, 0.1829833984375, -0.14453125, 0.142578125, 0.43994140625, 0.10357666015625, 0.1064453125, 0.01885986328125, -0.47705078125, 0.06884765625, -0.0689697265625, -0.0999755859375, 0.83251953125, 0.09033203125, -0.295166015625, 0.517578125, -0.66015625, 0.490966796875, -0.33984375, 0.58349609375, -0.1279296875, 0.1212158203125, 0.1221923828125, -1.001953125, -0.14404296875, 0.07647705078125, 0.1920166015625, -0.315185546875, -0.498779296875, 0.0234375, 1.1005859375, 0.412353515625, -0.6650390625, -0.262939453125, 0.7509765625, -0.1693115234375, -0.200439453125, -0.1273193359375, -0.398681640625, 0.19580078125, -0.1641845703125, -0.435302734375, 0.4716796875, 0.39990234375, -0.362060546875, -0.537109375, 0.12188720703125, 0.0657958984375, -0.10125732421875, 0.14111328125, -0.2276611328125, -0.0350341796875, 0.216552734375, 0.2093505859375, 0.413818359375, 0.39453125, 0.271240234375, -0.1434326171875, 0.1689453125, -0.0062255859375, 0.35205078125, 0.36181640625, 0.50439453125, -0.281005859375, -0.08856201171875, 0.1328125, -0.333740234375, -0.0126953125, -0.16455078125, 0.39306640625, 0.35498046875, -0.30029296875, 0.437744140625, -2.1328125, 0.5244140625, 0.440673828125, 0.1259765625, -0.5322265625, -0.133544921875, 0.0731201171875, -0.1435546875, 0.105224609375, 0.26171875, 0.349365234375, -0.12109375, 0.79052734375, -0.38818359375, 0.0589599609375, 0.66748046875, 0.2568359375, 0.2470703125, -0.19091796875, -0.372802734375, 0.2044677734375, 0.5107421875, 0.1549072265625, -0.0670166015625, -0.202392578125, 0.2412109375, 0.15576171875, 0.99609375, 0.272705078125, 0.307861328125, -0.17529296875, 0.29638671875, 0.150146484375, 0.0128173828125, 0.6220703125, -0.75537109375, -0.302734375, 0.5537109375, -0.60546875, 0.0521240234375, 0.75830078125, -0.45263671875, -0.25439453125, -0.160888671875, -0.41357421875, -0.207763671875, 0.12158203125, -0.9423828125, -0.08251953125, 0.1192626953125, 0.048095703125, 0.6357421875, -0.28173828125, 0.4111328125, -0.337158203125, -0.2249755859375, -0.1729736328125, -0.09814453125, 0.066650390625, -0.1361083984375, 0.1295166015625, 0.06317138671875, 0.1053466796875, 0.308349609375, -0.216064453125, -0.157958984375, -0.367919921875, 0.04217529296875, 0.45751953125, 1.265625, -0.1591796875, -0.309814453125, -0.5771484375, 0.5693359375, 0.345703125, 0.0283203125, 0.3330078125, -0.26708984375, -1.2880859375, -0.10015869140625, 0.0650634765625, -0.62158203125, 0.37109375, 0.59814453125, -0.06964111328125, 0.177001953125, -0.154541015625, 0.1065673828125, 0.13671875, 0.0064697265625, -0.48291015625, 0.300048828125, -0.1573486328125, -0.454345703125, 0.1541748046875, -0.13232421875, -0.30322265625, -0.212890625, -0.2197265625, -0.265380859375, -0.44091796875, 0.049072265625, -0.00634765625, -0.4306640625, -0.56591796875, -0.149658203125, -0.470703125, -0.09967041015625, 0.2119140625, -0.1270751953125, -0.36181640625, -0.232666015625, -0.172119140625, -0.21826171875, 0.252197265625, 0.4990234375, 0.19775390625, 0.2322998046875, -0.7109375, -0.326171875, -0.5576171875, -0.201904296875, 0.03741455078125, 0.22412109375, 0.149658203125, 0.06549072265625, 0.4384765625, -0.0103912353515625, 0.149658203125, -0.02484130859375, -0.189208984375, 0.4091796875, 0.258544921875, 0.55810546875, -0.16845703125, 0.04833984375, -0.791015625, -0.110107421875, 0.0535888671875, 0.1558837890625, -0.242431640625, -0.09576416015625, -0.52490234375, 0.30859375, -1.4208984375, 0.147216796875, -0.12371826171875, 1.259765625, -0.182373046875, 0.3818359375, 0.367431640625, 0.4404296875, -0.18603515625, -0.0738525390625, 0.283203125, -0.0809326171875, 0.37060546875, 0.587890625, -0.64501953125, -0.27685546875, -0.1309814453125, -0.18798828125, -0.0418701171875, 0.1363525390625, 0.468505859375, 0.05755615234375, -0.580078125, 0.1658935546875, -0.1865234375, -0.07659912109375, -0.7822265625, 0.88427734375, 0.0078125, -0.02001953125, -0.0126953125, -0.68408203125, 0.43115234375, 0.21484375, 0.474609375, -0.041015625, -0.62548828125, 0.71142578125, 0.27880859375, -0.10595703125, 1.095703125, 0.0753173828125, -0.0716552734375, -0.7841796875, -0.224609375, -0.2666015625, 0.002685546875, 0.173828125, 0.041778564453125, 0.4091796875, -0.19873046875, 0.12548828125, -0.01910400390625, -0.48388671875, -0.331298828125, 0.2763671875, -0.419921875, -0.2802734375, 0.10986328125, -0.393798828125, -0.1407470703125, -0.0977783203125, -0.05401611328125, 0.90185546875, -0.0472412109375, -0.294921875, -0.09619140625, 0.1707763671875, 0.092529296875, -0.0467529296875, 0.298583984375, -0.12890625, 0.275390625, 0.587890625, -0.4775390625, 0.205810546875, -0.2724609375, -0.83056640625, 0.1490478515625, 0.2626953125, 0.0751953125, -0.04132080078125, 0.0672607421875, 0.15380859375, 0.0914306640625, -0.22412109375, 0.44775390625, 0.7060546875]}, {"frame_number": 8, "vector": [-0.143798828125, 0.3291015625, 0.008056640625, -0.1939697265625, -0.62158203125, 0.35302734375, -0.0089111328125, -0.46923828125, -0.1351318359375, -0.035003662109375, -0.14892578125, -0.00830078125, 0.35595703125, 0.194091796875, -0.040283203125, 0.306640625, 0.1173095703125, -0.093994140625, -0.094970703125, -0.4521484375, -0.57666015625, -6.21875, -0.35986328125, -0.69189453125, 0.00762939453125, -0.0888671875, -0.14599609375, -0.7451171875, 0.466796875, -0.186767578125, -0.9658203125, -0.6591796875, -0.72607421875, 0.02239990234375, -0.08203125, 0.083251953125, 0.293212890625, -0.6591796875, -0.07623291015625, -0.4482421875, -0.07647705078125, 0.20068359375, 0.226806640625, 0.052001953125, 0.91552734375, -0.34814453125, -0.394775390625, -0.2152099609375, -0.5732421875, -0.1689453125, 0.047027587890625, -0.259521484375, 0.07965087890625, 0.51318359375, 0.24658203125, 0.0477294921875, 0.407958984375, -0.2763671875, 0.286865234375, -0.0242919921875, -0.12548828125, 0.21728515625, 0.3046875, 0.1807861328125, -0.231201171875, -0.31298828125, -0.50048828125, -0.07177734375, 0.1627197265625, 0.0975341796875, 0.9775390625, -0.0841064453125, 0.028076171875, -0.0186767578125, -0.443115234375, 0.8427734375, -0.36474609375, -0.47216796875, 0.19189453125, -0.65966796875, 1.15625, 0.321533203125, 0.59375, 0.52001953125, -0.27490234375, 0.0406494140625, 0.1724853515625, 0.02783203125, 0.119140625, -0.59765625, -0.185302734375, -0.11956787109375, 0.044189453125, -0.155517578125, 0.26513671875, 0.320068359375, 0.070068359375, 0.5615234375, 0.0728759765625, -0.66455078125, 0.7392578125, 0.31884765625, -0.361328125, -0.62548828125, -0.00830078125, -0.162841796875, -0.00360107421875, -0.160888671875, -0.172607421875, -0.054931640625, -0.083984375, 0.166259765625, 0.087646484375, -0.47265625, 0.1607666015625, -0.56640625, -0.1865234375, -0.1248779296875, 0.30908203125, 0.52392578125, 0.037109375, -0.210693359375, -0.053466796875, -0.72802734375, -0.0826416015625, 0.208984375, -0.1527099609375, 0.347900390625, 0.1932373046875, 0.3837890625, -0.005859375, -0.392822265625, 0.0230712890625, 0.0765380859375, 0.556640625, -0.0303955078125, 0.0460205078125, 0.77734375, 0.6083984375, 0.010986328125, 0.0755615234375, 0.28173828125, 0.181396484375, -0.0745849609375, -0.25732421875, 0.1016845703125, -0.00677490234375, -0.343017578125, 0.22119140625, 0.99560546875, -0.1522216796875, -1.5595703125, 0.0712890625, -0.41552734375, 0.444091796875, -0.243408203125, -0.287109375, 0.93359375, 0.30322265625, -0.42529296875, -0.2471923828125, 0.3203125, 0.491455078125, 0.29638671875, -0.32421875, -0.27587890625, -0.003662109375, 0.67724609375, 0.033447265625, 0.0914306640625, 0.1343994140625, 0.4580078125, -0.1650390625, 0.66455078125, -0.292236328125, 1.044921875, -1.134765625, 0.0743408203125, -0.0078125, -0.108642578125, 0.038818359375, 0.19775390625, -0.1837158203125, 0.036376953125, 0.0635986328125, -0.0103759765625, 0.015869140625, 0.173828125, -0.09130859375, 0.1824951171875, 0.1827392578125, 0.1614990234375, 0.445556640625, -0.361328125, 0.529296875, -0.2110595703125, 0.431396484375, -0.245849609375, -0.39599609375, 1.1630859375, 0.463623046875, -0.06591796875, -0.2398681640625, 0.0606689453125, -0.02789306640625, 0.06787109375, 0.66064453125, 0.0120849609375, -0.66796875, -0.142578125, 0.2037353515625, -0.11328125, 0.288330078125, -0.82470703125, -0.5380859375, -0.4873046875, 0.4736328125, 1.056640625, -0.0047454833984375, -0.395263671875, -0.427001953125, -0.58251953125, -0.17236328125, -0.60498046875, -0.167724609375, 0.165771484375, -0.11328125, -0.23486328125, 0.76806640625, 0.0042724609375, -0.11175537109375, 0.27197265625, 0.0576171875, 0.265380859375, -0.3212890625, -0.546875, 0.78955078125, -0.0516357421875, -0.92626953125, -0.101806640625, 0.70703125, 0.2337646484375, 0.03076171875, 0.098388671875, -0.16650390625, 1.029296875, -0.1734619140625, 0.01513671875, -0.44140625, -0.075439453125, 0.00390625, -0.0330810546875, 0.3291015625, -0.045166015625, -0.7001953125, -0.095947265625, -0.185302734375, -0.0028076171875, 0.497802734375, 0.49462890625, -0.01544189453125, -0.1800537109375, -0.062744140625, -0.015869140625, -0.286376953125, 0.01318359375, -0.8056640625, -0.29931640625, 0.5458984375, -0.45751953125, 0.04022216796875, 0.404296875, 0.32666015625, 0.257568359375, -0.1964111328125, -0.6845703125, -0.05816650390625, 0.12457275390625, -0.349853515625, -0.419189453125, -0.313232421875, 0.662109375, -0.41357421875, 0.122314453125, 0.407958984375, 0.325439453125, -0.133056640625, 0.5361328125, -0.5888671875, 1.12109375, -0.019256591796875, 0.0582275390625, 0.0396728515625, -0.37744140625, 0.06884765625, 0.1873779296875, 0.175048828125, 0.376953125, 0.6396484375, -0.256103515625, -0.0274200439453125, 0.76123046875, -0.0229949951171875, -0.0458984375, -0.290283203125, 0.36279296875, 0.327880859375, -0.70361328125, 0.1490478515625, -0.07275390625, 0.372314453125, -0.33251953125, -0.01416015625, -0.31103515625, -0.34423828125, 0.387451171875, -0.03338623046875, 0.012451171875, -0.2763671875, 0.196533203125, -0.912109375, 0.12548828125, 0.098388671875, -0.4697265625, -0.87158203125, 0.6728515625, -0.435791015625, 0.32666015625, -0.60888671875, -0.00543212890625, -0.5576171875, -0.031982421875, -0.17333984375, -0.5048828125, 0.208740234375, 0.223876953125, 0.0850830078125, -0.603515625, 0.04278564453125, -0.00537109375, -0.69775390625, 0.02490234375, -0.277099609375, -0.7021484375, -0.509765625, 0.217529296875, -0.474609375, 0.3984375, 0.2255859375, 0.0313720703125, -0.328125, -0.2783203125, 0.28466796875, 0.1552734375, 0.420166015625, 0.318603515625, 0.2509765625, -0.1668701171875, -0.349609375, -0.2156982421875, -0.1669921875, 0.05810546875, 0.4990234375, 0.286376953125, -0.6103515625, -0.43359375, -0.03955078125, 0.2279052734375, 0.07684326171875, -0.0770263671875, -0.044921875, 0.25830078125, -0.1796875, 0.3359375, 0.283935546875, -0.0704345703125, 0.43310546875, -0.25146484375, 0.169921875, -0.53369140625, -0.063720703125, -0.381591796875, 0.12841796875, 0.71044921875, 0.23681640625, -0.0377197265625, -0.0364990234375, -0.11767578125, -0.27734375, 0.52734375, -0.650390625, -0.224365234375, 0.301025390625, 0.270751953125, 1.037109375, 0.6630859375, -0.118408203125, -0.6337890625, -0.0196533203125, -0.4365234375, 0.232666015625, 0.1513671875, -0.104736328125, -0.466796875, -0.195068359375, 0.0223388671875, -0.393310546875, 0.04669189453125, 0.36962890625, 0.0042724609375, 0.190185546875, -0.85791015625, -0.74267578125, -0.3251953125, -0.392578125, 0.57177734375, 0.7158203125, -0.29833984375, 1.037109375, -0.237548828125, 0.1236572265625, -0.02520751953125, -0.224853515625, -0.0948486328125, -0.013427734375, -0.50537109375, -1.40625, -0.1539306640625, 0.1026611328125, 0.1903076171875, 0.06683349609375, -6.21875, -0.41357421875, -0.052001953125, -0.79248046875, -0.1922607421875, -0.1341552734375, -1.20703125, -0.62109375, -0.231201171875, -0.003662109375, 0.6220703125, 0.63818359375, -0.28271484375, -0.0391845703125, 0.205322265625, 0.234130859375, 0.016845703125, -0.0482177734375, 0.173828125, -0.006591796875, 0.180908203125, -0.11090087890625, 0.0762939453125, 0.416259765625, -0.279541015625, -0.165283203125, 0.4189453125, -0.6728515625, -0.623046875, -0.078369140625, 0.12939453125, -0.2646484375, -0.332763671875, -0.4892578125, -0.0283203125, 0.55322265625, 0.76220703125, -0.35400390625, -0.404296875, 0.215087890625, -1.0234375, 0.325927734375, -0.0184326171875, -0.0210113525390625, 0.220947265625, 0.3125, -0.1014404296875, -0.7763671875, -0.77099609375, 0.12469482421875, 0.0531005859375, -1.09765625, -0.0166015625, 0.349365234375, -0.0164794921875, 0.235595703125, -0.08782958984375, 0.1162109375, 0.180419921875, 0.209716796875, 0.57666015625, 0.339111328125, 0.429931640625, -0.48291015625, 0.017333984375, 0.07720947265625, 0.0155029296875, 0.1025390625, 0.075439453125, -0.0810546875, 0.05413818359375, 0.6611328125, -0.171875, -0.24462890625, -0.1259765625, -0.0343017578125, 0.1405029296875, -0.19921875, 0.05126953125, 0.215087890625, 0.018798828125, 0.1998291015625, 0.414794921875, -0.208740234375, -0.14013671875, -0.10595703125, -0.333740234375, -0.58349609375, -0.1298828125, 0.7138671875, 0.2080078125, -0.1763916015625, 0.05987548828125, 0.2337646484375, 1.2880859375, 0.148681640625, 0.14501953125, 0.128173828125, 0.1103515625, -0.311279296875, -0.497802734375, 0.009521484375, -0.033935546875, 0.362548828125, 0.298828125, 0.06689453125, -0.0413818359375, 0.40283203125, -0.09521484375, 0.7373046875, 0.13916015625, 0.361083984375, 0.5205078125, -0.212158203125, 0.4833984375, -1.541015625, -0.36767578125, -0.1864013671875, 0.1199951171875, -0.0072021484375, -0.341796875, 0.74462890625, 0.281494140625, -0.1220703125, 0.0732421875, 0.296875, -0.0621337890625, -0.24267578125, -0.92578125, -0.448486328125, -0.313720703125, -0.453857421875, -0.958984375, 0.261962890625, 0.299560546875, 0.5380859375, -0.3388671875, -0.09942626953125, 0.0433349609375, -0.042236328125, 0.376953125, -0.2265625, 0.116455078125, 0.179931640625, 0.322509765625, -0.3369140625, 0.208251953125, -0.308349609375, -0.484619140625, -0.314453125, 0.09716796875, -0.148193359375, -0.13330078125, -0.01995849609375, -0.10650634765625, 0.02001953125, 0.45849609375, -0.339599609375, 0.206298828125, 0.385009765625, 0.0066375732421875, -0.220947265625, -0.0146484375, 0.08209228515625, 0.0736083984375, -0.059814453125, -0.20458984375, -1.458984375, -0.427490234375, 0.16259765625, -0.0712890625, -0.31298828125, -0.381591796875, -0.224365234375, -0.143310546875, -0.0745849609375, -0.28125, -0.262939453125, -0.89697265625, 0.252685546875, 0.40478515625, -0.2037353515625, -0.20947265625, -0.123291015625, -0.55859375, 0.248046875, 0.2457275390625, -0.041534423828125, -0.254150390625, 0.03228759765625, -0.083984375, 0.0716552734375, 0.0692138671875, -0.23974609375, -0.0328369140625, 0.07354736328125, 0.3466796875, 0.22802734375, -0.0748291015625, 0.158447265625, 0.03704833984375, -0.3828125, 0.6962890625, 0.04541015625, -0.11083984375, -0.31103515625, 1.30078125, 0.075927734375, -0.25634765625, 0.02880859375, -0.50341796875, 0.357421875, -0.2384033203125, -0.181396484375, 0.32373046875, 0.1673583984375, -0.012908935546875, 0.292724609375, 0.09674072265625, 0.246826171875, -0.35693359375, 0.3837890625, 0.2421875, -0.62841796875, 0.135009765625, 0.278076171875, -0.10400390625, -0.2705078125, -0.26708984375, -0.31396484375, -0.8203125, 0.0279541015625, -0.10906982421875, -0.07452392578125, -0.250244140625, 0.09228515625, -0.05615234375, 0.48095703125, -0.457275390625, -1.365234375, -0.07861328125, -0.13720703125, 0.209228515625, 0.2337646484375, -0.0592041015625, 0.362548828125, 0.158447265625, -0.1875, 0.238037109375, -0.02630615234375, 0.16796875, 0.4267578125, 0.020263671875, 0.146484375, -0.219482421875, 0.29296875, 0.1968994140625, -0.38427734375, -0.6064453125, -0.20654296875, -0.44873046875, 0.057861328125, 0.23974609375, -0.290283203125, 0.7861328125, -0.0655517578125, 0.1041259765625, -0.86962890625, 0.000640869140625, -0.004150390625, 0.19970703125, -0.61865234375, 0.309326171875, -0.5361328125, 0.37939453125, -0.031982421875, 0.208984375, -0.138671875, -0.513671875, -0.6484375, 0.1064453125, -0.413818359375, 0.43359375, -0.327880859375, 0.9521484375, -0.33251953125, 0.0531005859375, -0.122314453125, -0.010498046875, -0.419921875, 0.09765625, 1.3671875, 0.028900146484375, 0.2509765625, -0.088623046875, 0.38916015625, -1.076171875, -0.245849609375, 0.51171875, 0.0382080078125, -0.25927734375, 0.4990234375, 0.138671875, 0.061767578125, 0.203125, 0.06500244140625, 0.18603515625, -0.1234130859375, 0.46826171875, -0.515625, 0.04833984375, -0.630859375, 0.04443359375, 0.09130859375, -0.45703125, -0.00030517578125, -0.060791015625, -0.1002197265625, 0.08251953125, -0.90234375, -0.24462890625, 0.27392578125, -0.1260986328125, 0.2420654296875, -0.025848388671875, -0.09765625, -0.8349609375, 0.0303955078125, 0.294189453125, 0.00543212890625, 0.0850830078125, -0.2393798828125, -0.533203125, 0.60693359375, -0.401611328125, -0.037384033203125, -0.57763671875, 0.00433349609375, -0.384765625, 0.171630859375, -0.015625, 0.065673828125, 0.0654296875, 0.1810302734375, -2.2890625, 0.3662109375, -0.183349609375, 0.264892578125, 0.6416015625, -0.0303955078125, -0.3125, -0.23828125, 0.17333984375, -0.1025390625, 0.208740234375, -0.0404052734375, 0.0443115234375, 0.45361328125, 0.4599609375, 0.307861328125, 0.006103515625, 0.2529296875, 0.6337890625, -0.2724609375, 0.3876953125, 0.10107421875, 0.112548828125, -0.2587890625, 0.095703125, 0.26416015625, 0.2705078125, -0.7216796875, 0.5712890625, 0.1182861328125, -0.173583984375, -0.0076904296875, -0.49365234375, 0.011474609375, -0.0355224609375, 0.32666015625, -0.28955078125, -0.022979736328125, -0.204833984375, -0.22021484375, 0.384765625, -0.048095703125, -0.278076171875, -0.646484375, -0.410888671875, 0.283935546875, 0.1029052734375, -0.18017578125, 0.357666015625, -0.1412353515625, 0.311767578125, -0.425537109375, -0.29833984375, 0.1964111328125, -0.325439453125, -0.478759765625, -0.6611328125, -0.63134765625, -0.028564453125, 0.7900390625, 0.60400390625, -0.28564453125, 0.0103759765625, 0.68359375, -0.68359375, 0.236083984375, -0.6572265625, -0.6337890625, -0.49853515625, 0.14208984375, 0.08282470703125, -0.337646484375, -0.32470703125, 0.1602783203125, -0.1103515625, 0.15478515625, 0.438720703125, 0.103759765625, 0.1248779296875, 0.0115203857421875, -0.478515625, 0.062255859375, -0.063720703125, -0.0931396484375, 0.83642578125, 0.093994140625, -0.31005859375, 0.51416015625, -0.65576171875, 0.4921875, -0.333251953125, 0.5712890625, -0.134521484375, 0.10498046875, 0.11474609375, -0.998046875, -0.14794921875, 0.0615234375, 0.171630859375, -0.3154296875, -0.48583984375, 0.028564453125, 1.138671875, 0.39306640625, -0.69921875, -0.268798828125, 0.7646484375, -0.1900634765625, -0.215087890625, -0.1402587890625, -0.393798828125, 0.18798828125, -0.1822509765625, -0.427001953125, 0.4599609375, 0.4013671875, -0.362548828125, -0.53564453125, 0.10858154296875, 0.0640869140625, -0.156494140625, 0.14501953125, -0.2012939453125, -0.040771484375, 0.233154296875, 0.2066650390625, 0.405517578125, 0.3798828125, 0.266357421875, -0.115234375, 0.15625, 0.004486083984375, 0.36181640625, 0.3447265625, 0.50390625, -0.308349609375, -0.0789794921875, 0.141357421875, -0.325439453125, -0.01318359375, -0.1781005859375, 0.3896484375, 0.379638671875, -0.288818359375, 0.449462890625, -2.14453125, 0.5166015625, 0.43603515625, 0.138671875, -0.58837890625, -0.1297607421875, 0.0625, -0.144775390625, 0.120849609375, 0.22607421875, 0.351318359375, -0.118896484375, 0.7509765625, -0.381591796875, 0.09881591796875, 0.6943359375, 0.25146484375, 0.240234375, -0.220703125, -0.37109375, 0.1912841796875, 0.53955078125, 0.1585693359375, -0.05810546875, -0.187255859375, 0.251953125, 0.1265869140625, 0.98046875, 0.28125, 0.32373046875, -0.194580078125, 0.303466796875, 0.1279296875, 0.00567626953125, 0.65869140625, -0.7197265625, -0.28857421875, 0.53662109375, -0.61376953125, 0.0474853515625, 0.71533203125, -0.4443359375, -0.24462890625, -0.147705078125, -0.4130859375, -0.221435546875, 0.123779296875, -0.8994140625, -0.09326171875, 0.177001953125, 0.0018310546875, 0.61962890625, -0.25, 0.41845703125, -0.2900390625, -0.2291259765625, -0.1834716796875, -0.082275390625, 0.064697265625, -0.1358642578125, 0.1478271484375, 0.05902099609375, 0.1427001953125, 0.321044921875, -0.196044921875, -0.148681640625, -0.3916015625, 0.04364013671875, 0.45947265625, 1.22265625, -0.126708984375, -0.29150390625, -0.60302734375, 0.61328125, 0.32861328125, 0.0419921875, 0.33837890625, -0.255859375, -1.3369140625, -0.11651611328125, 0.076904296875, -0.61474609375, 0.377685546875, 0.57177734375, -0.06689453125, 0.184814453125, -0.12493896484375, 0.0748291015625, 0.102294921875, 0.00555419921875, -0.47900390625, 0.283447265625, -0.1519775390625, -0.48193359375, 0.200927734375, -0.1151123046875, -0.31884765625, -0.197265625, -0.222900390625, -0.253662109375, -0.450927734375, 0.045654296875, 0.017333984375, -0.41015625, -0.55908203125, -0.1435546875, -0.49658203125, -0.12188720703125, 0.203857421875, -0.083740234375, -0.3759765625, -0.234130859375, -0.18896484375, -0.236572265625, 0.187744140625, 0.52685546875, 0.1455078125, 0.212646484375, -0.7138671875, -0.313720703125, -0.55322265625, -0.232666015625, 0.032958984375, 0.23046875, 0.17333984375, 0.07293701171875, 0.4296875, -0.0285491943359375, 0.1295166015625, 0.028045654296875, -0.14892578125, 0.424560546875, 0.232421875, 0.578125, -0.162841796875, 0.0302734375, -0.75439453125, -0.1309814453125, 0.031982421875, 0.173095703125, -0.206298828125, -0.080322265625, -0.5068359375, 0.3076171875, -1.455078125, 0.1146240234375, -0.11627197265625, 1.2509765625, -0.1737060546875, 0.378662109375, 0.35595703125, 0.46337890625, -0.2017822265625, -0.0740966796875, 0.2587890625, -0.047119140625, 0.39208984375, 0.5791015625, -0.6298828125, -0.25439453125, -0.1337890625, -0.1915283203125, -0.042938232421875, 0.152099609375, 0.48828125, 0.03826904296875, -0.58203125, 0.15283203125, -0.178466796875, -0.05780029296875, -0.7802734375, 0.85107421875, -0.026611328125, -0.0205078125, -0.0054931640625, -0.6669921875, 0.429443359375, 0.218505859375, 0.46533203125, -0.0517578125, -0.638671875, 0.68408203125, 0.2529296875, -0.1207275390625, 1.091796875, 0.06005859375, -0.060546875, -0.77734375, -0.2412109375, -0.281982421875, 0.008056640625, 0.17529296875, 0.05755615234375, 0.404296875, -0.186279296875, 0.1365966796875, -0.02239990234375, -0.469970703125, -0.308837890625, 0.282958984375, -0.42626953125, -0.2998046875, 0.0836181640625, -0.415283203125, -0.127685546875, -0.10467529296875, -0.03485107421875, 0.91357421875, -0.0284423828125, -0.290771484375, -0.1097412109375, 0.17138671875, 0.035247802734375, -0.015869140625, 0.261962890625, -0.11370849609375, 0.27294921875, 0.5703125, -0.48876953125, 0.22021484375, -0.2509765625, -0.83154296875, 0.1907958984375, 0.261962890625, 0.07080078125, -0.01885986328125, 0.078857421875, 0.147705078125, 0.064208984375, -0.2265625, 0.482666015625, 0.6943359375]}, {"frame_number": 9, "vector": [-0.1688232421875, 0.300048828125, 0.0029296875, -0.1583251953125, -0.62646484375, 0.32177734375, 0.01934814453125, -0.48291015625, -0.1097412109375, -0.017425537109375, -0.135009765625, 0.09716796875, 0.330078125, 0.194580078125, -0.0560302734375, 0.2958984375, 0.135498046875, -0.14111328125, -0.08758544921875, -0.48291015625, -0.5556640625, -6.109375, -0.3671875, -0.728515625, 0.01971435546875, -0.0594482421875, -0.18359375, -0.8017578125, 0.517578125, -0.18603515625, -0.96484375, -0.61376953125, -0.67578125, -0.01171875, -0.10400390625, -0.00555419921875, 0.327392578125, -0.66015625, -0.0694580078125, -0.390380859375, -0.07794189453125, 0.1942138671875, 0.2449951171875, 0.0469970703125, 0.91455078125, -0.407470703125, -0.397705078125, -0.21337890625, -0.59619140625, -0.1689453125, 0.02197265625, -0.26025390625, 0.16064453125, 0.5478515625, 0.256591796875, -0.0322265625, 0.41748046875, -0.218017578125, 0.24560546875, -0.02105712890625, -0.0771484375, 0.216064453125, 0.276611328125, 0.1690673828125, -0.215087890625, -0.349609375, -0.50390625, -0.051025390625, 0.195068359375, 0.0125732421875, 0.9609375, -0.04443359375, -0.019775390625, -0.097412109375, -0.439453125, 0.8798828125, -0.323974609375, -0.46875, 0.2685546875, -0.66943359375, 1.189453125, 0.274658203125, 0.5791015625, 0.51025390625, -0.283203125, 0.0408935546875, 0.1624755859375, -0.027099609375, 0.16455078125, -0.57470703125, -0.1781005859375, -0.137939453125, 0.0465087890625, -0.10595703125, 0.256591796875, 0.35595703125, 0.0970458984375, 0.533203125, 0.038543701171875, -0.6806640625, 0.73046875, 0.368896484375, -0.373046875, -0.62451171875, -0.0100250244140625, -0.0950927734375, 0.00811767578125, -0.111083984375, -0.177978515625, -0.022216796875, -0.08612060546875, 0.1806640625, 0.10089111328125, -0.447509765625, 0.12744140625, -0.6025390625, -0.2255859375, -0.1405029296875, 0.338134765625, 0.48095703125, 0.05126953125, -0.260498046875, -0.0380859375, -0.7138671875, -0.0723876953125, 0.25048828125, -0.1025390625, 0.355712890625, 0.1934814453125, 0.335693359375, -0.0194091796875, -0.376953125, 0.02044677734375, 0.01837158203125, 0.5595703125, -0.03125, 0.02862548828125, 0.7509765625, 0.56005859375, -0.00439453125, 0.0242919921875, 0.283935546875, 0.1981201171875, -0.11083984375, -0.28076171875, 0.101806640625, -0.014129638671875, -0.328369140625, 0.2049560546875, 1.009765625, -0.194580078125, -1.541015625, 0.065185546875, -0.3583984375, 0.416015625, -0.231201171875, -0.2459716796875, 0.93701171875, 0.239013671875, -0.41455078125, -0.264892578125, 0.264404296875, 0.46240234375, 0.2841796875, -0.3173828125, -0.275390625, 0.0216064453125, 0.623046875, 0.0384521484375, 0.059814453125, 0.168212890625, 0.39306640625, -0.133544921875, 0.6494140625, -0.23974609375, 0.9755859375, -1.166015625, 0.17724609375, -0.0648193359375, -0.087890625, 0.00830078125, 0.21044921875, -0.16455078125, 0.022247314453125, 0.0565185546875, -0.0205078125, 0.05322265625, 0.198974609375, -0.0712890625, 0.167236328125, 0.166748046875, 0.1053466796875, 0.432861328125, -0.37158203125, 0.5439453125, -0.18603515625, 0.444091796875, -0.210205078125, -0.292236328125, 1.162109375, 0.446533203125, -0.0166015625, -0.2364501953125, 0.06396484375, 0.047119140625, 0.04248046875, 0.638671875, 0.04022216796875, -0.72216796875, -0.1380615234375, 0.1937255859375, -0.13037109375, 0.24560546875, -0.830078125, -0.473876953125, -0.515625, 0.50830078125, 1.044921875, 0.0055389404296875, -0.442138671875, -0.4326171875, -0.5263671875, -0.116943359375, -0.57177734375, -0.197509765625, 0.18603515625, -0.1158447265625, -0.26318359375, 0.76904296875, 0.0048828125, -0.1663818359375, 0.2529296875, 0.048828125, 0.230712890625, -0.32177734375, -0.5263671875, 0.77197265625, -0.10052490234375, -0.9990234375, -0.05517578125, 0.71728515625, 0.256591796875, -0.0230712890625, 0.1484375, -0.211181640625, 0.99609375, -0.2236328125, 0.01708984375, -0.38720703125, -0.151123046875, 0.00048828125, -0.03271484375, 0.2666015625, -0.08837890625, -0.67578125, -0.087890625, -0.137939453125, -0.02734375, 0.452392578125, 0.4755859375, -0.00604248046875, -0.21484375, -0.072509765625, 0.01025390625, -0.238037109375, 0.051513671875, -0.78955078125, -0.31884765625, 0.552734375, -0.4228515625, 0.06201171875, 0.378173828125, 0.323486328125, 0.28125, -0.27197265625, -0.70361328125, -0.08929443359375, 0.1683349609375, -0.35888671875, -0.403076171875, -0.293701171875, 0.68408203125, -0.45703125, 0.11474609375, 0.370849609375, 0.302734375, -0.1187744140625, 0.5244140625, -0.5283203125, 1.15625, 0.00244140625, 0.030609130859375, 0.0770263671875, -0.37109375, 0.0364990234375, 0.1197509765625, 0.185302734375, 0.405029296875, 0.6015625, -0.237548828125, -0.00537109375, 0.7568359375, 0.02606201171875, -0.0234375, -0.26416015625, 0.355224609375, 0.327880859375, -0.70556640625, 0.13037109375, -0.1512451171875, 0.345703125, -0.400146484375, -0.0068359375, -0.31884765625, -0.318115234375, 0.40380859375, -0.0675048828125, 0.0123291015625, -0.28759765625, 0.223876953125, -0.931640625, 0.1278076171875, 0.0482177734375, -0.47509765625, -0.84130859375, 0.61962890625, -0.4560546875, 0.3232421875, -0.59228515625, 0.013427734375, -0.51416015625, -0.0816650390625, -0.177978515625, -0.464111328125, 0.1912841796875, 0.2054443359375, 0.1470947265625, -0.6484375, 0.015380859375, -0.0128173828125, -0.70556640625, 0.03594970703125, -0.32470703125, -0.7177734375, -0.533203125, 0.1796875, -0.45947265625, 0.298583984375, 0.2425537109375, 0.0465087890625, -0.33544921875, -0.3466796875, 0.234130859375, 0.130859375, 0.411376953125, 0.35107421875, 0.2430419921875, -0.155517578125, -0.349609375, -0.2210693359375, -0.189453125, 0.019775390625, 0.52001953125, 0.328369140625, -0.619140625, -0.4462890625, -0.063232421875, 0.2481689453125, 0.0482177734375, -0.1151123046875, -0.0057373046875, 0.265625, -0.161865234375, 0.29833984375, 0.310791015625, -0.0487060546875, 0.44287109375, -0.209228515625, 0.181640625, -0.5400390625, -0.0751953125, -0.360595703125, 0.143798828125, 0.7060546875, 0.1876220703125, -0.0809326171875, 0.00030517578125, -0.15869140625, -0.286376953125, 0.5595703125, -0.6474609375, -0.22802734375, 0.27392578125, 0.2236328125, 1.068359375, 0.658203125, -0.161865234375, -0.62548828125, -0.0631103515625, -0.41455078125, 0.18408203125, 0.12841796875, -0.1083984375, -0.44091796875, -0.144287109375, 0.015625, -0.334716796875, 0.061492919921875, 0.3828125, -0.01171875, 0.1953125, -0.888671875, -0.72900390625, -0.31103515625, -0.397705078125, 0.572265625, 0.7392578125, -0.308837890625, 1.1015625, -0.19970703125, 0.1629638671875, -0.01654052734375, -0.179443359375, -0.092529296875, -0.0576171875, -0.51611328125, -1.4140625, -0.06842041015625, 0.1187744140625, 0.1973876953125, 0.0821533203125, -6.109375, -0.41455078125, -0.031982421875, -0.79541015625, -0.2269287109375, -0.0712890625, -1.205078125, -0.6513671875, -0.2744140625, -0.02825927734375, 0.55859375, 0.63671875, -0.211669921875, 0.02386474609375, 0.2548828125, 0.255615234375, 0.0159912109375, -0.03857421875, 0.208251953125, 0.0498046875, 0.18505859375, -0.11187744140625, 0.100341796875, 0.460693359375, -0.30029296875, -0.22021484375, 0.38818359375, -0.6962890625, -0.64697265625, -0.070556640625, 0.086181640625, -0.307861328125, -0.319091796875, -0.458984375, -0.0286865234375, 0.57470703125, 0.77099609375, -0.35546875, -0.3505859375, 0.2235107421875, -1.00390625, 0.31298828125, -0.0789794921875, -0.023529052734375, 0.1961669921875, 0.4140625, -0.063720703125, -0.810546875, -0.7451171875, 0.146240234375, 0.084228515625, -1.0703125, -0.072265625, 0.36474609375, -0.0081787109375, 0.270751953125, -0.06231689453125, 0.12371826171875, 0.18212890625, 0.1162109375, 0.5185546875, 0.314697265625, 0.43701171875, -0.369384765625, 0.038330078125, 0.052978515625, 0.0142822265625, 0.104736328125, 0.1004638671875, -0.0565185546875, 0.0816650390625, 0.6474609375, -0.161376953125, -0.25439453125, -0.12890625, -0.065185546875, 0.0828857421875, -0.2197265625, 0.083251953125, 0.204345703125, 0.0330810546875, 0.15380859375, 0.3564453125, -0.24658203125, -0.1829833984375, -0.10107421875, -0.35205078125, -0.6171875, -0.0950927734375, 0.72509765625, 0.18798828125, -0.1866455078125, 0.071533203125, 0.236572265625, 1.271484375, 0.17919921875, 0.10595703125, 0.1240234375, 0.07275390625, -0.35400390625, -0.477294921875, -0.04248046875, -0.06005859375, 0.317626953125, 0.28515625, 0.11328125, -0.03485107421875, 0.412109375, -0.0526123046875, 0.8056640625, 0.135986328125, 0.36376953125, 0.52001953125, -0.21728515625, 0.494384765625, -1.4873046875, -0.44677734375, -0.191650390625, 0.0955810546875, -0.0244140625, -0.375732421875, 0.775390625, 0.29052734375, -0.098388671875, 0.1123046875, 0.31982421875, -0.0953369140625, -0.20361328125, -0.91455078125, -0.409423828125, -0.263916015625, -0.4541015625, -0.9609375, 0.247802734375, 0.33203125, 0.517578125, -0.39013671875, -0.093017578125, 0.0804443359375, -0.0682373046875, 0.3583984375, -0.25390625, 0.1156005859375, 0.189453125, 0.387939453125, -0.39990234375, 0.20068359375, -0.314697265625, -0.482666015625, -0.322021484375, 0.171875, -0.173583984375, -0.120361328125, 0.0518798828125, -0.134033203125, -0.01025390625, 0.490234375, -0.319091796875, 0.1904296875, 0.36083984375, -0.0203857421875, -0.230224609375, 0.0262451171875, 0.113525390625, 0.03076171875, -0.114990234375, -0.173095703125, -1.48828125, -0.425537109375, 0.191162109375, -0.04083251953125, -0.350830078125, -0.33935546875, -0.27490234375, -0.0859375, -0.042816162109375, -0.253662109375, -0.2200927734375, -0.86083984375, 0.1630859375, 0.337890625, -0.2000732421875, -0.197265625, -0.148681640625, -0.57373046875, 0.265625, 0.220458984375, -0.03448486328125, -0.28955078125, 0.00701904296875, -0.0439453125, 0.0103759765625, 0.0755615234375, -0.238037109375, 0.002471923828125, 0.03057861328125, 0.39111328125, 0.234619140625, -0.0516357421875, 0.191162109375, -0.0137939453125, -0.36279296875, 0.67431640625, 0.0352783203125, -0.154296875, -0.302490234375, 1.322265625, 0.0537109375, -0.29296875, 0.011962890625, -0.377685546875, 0.351318359375, -0.2490234375, -0.19384765625, 0.37841796875, 0.1820068359375, -0.00830078125, 0.2578125, 0.1396484375, 0.2734375, -0.34912109375, 0.376220703125, 0.198486328125, -0.650390625, 0.174560546875, 0.3427734375, -0.099609375, -0.26416015625, -0.2271728515625, -0.2958984375, -0.8271484375, 0.0189208984375, -0.1142578125, -0.12939453125, -0.263671875, 0.071044921875, -0.119140625, 0.4560546875, -0.4560546875, -1.37890625, -0.0933837890625, -0.0830078125, 0.216064453125, 0.232666015625, -0.114501953125, 0.29541015625, 0.1591796875, -0.1474609375, 0.271728515625, -0.0146484375, 0.177001953125, 0.443603515625, 0.030029296875, 0.14990234375, -0.177734375, 0.291748046875, 0.227294921875, -0.3623046875, -0.61865234375, -0.18798828125, -0.4208984375, 0.00537109375, 0.283203125, -0.309326171875, 0.82080078125, -0.037109375, 0.1014404296875, -0.80712890625, 0.0186767578125, 0.0079345703125, 0.234130859375, -0.5771484375, 0.27294921875, -0.58203125, 0.41845703125, -0.04522705078125, 0.223388671875, -0.11553955078125, -0.483154296875, -0.73583984375, 0.07415771484375, -0.420654296875, 0.45458984375, -0.319091796875, 0.9931640625, -0.318359375, 0.0751953125, -0.102294921875, -0.0372314453125, -0.4521484375, 0.11181640625, 1.4111328125, 0.000732421875, 0.264892578125, -0.056396484375, 0.409423828125, -1.04296875, -0.212890625, 0.5244140625, 0.053466796875, -0.24658203125, 0.5478515625, 0.2003173828125, 0.11767578125, 0.13671875, 0.05743408203125, 0.171630859375, -0.068359375, 0.4375, -0.439697265625, 0.00439453125, -0.66455078125, 0.087158203125, 0.0909423828125, -0.4794921875, -0.00091552734375, -0.05010986328125, -0.161376953125, 0.138427734375, -0.83935546875, -0.28857421875, 0.2354736328125, -0.10821533203125, 0.1959228515625, -0.050262451171875, -0.149658203125, -0.77099609375, 0.019775390625, 0.2763671875, -0.0294189453125, 0.1427001953125, -0.25390625, -0.5244140625, 0.58154296875, -0.434326171875, 0.0081787109375, -0.6259765625, 0.00372314453125, -0.32421875, 0.126708984375, 0.03271484375, 0.0379638671875, 0.0924072265625, 0.1956787109375, -2.3203125, 0.34521484375, -0.25390625, 0.298095703125, 0.69677734375, -0.068359375, -0.32275390625, -0.2568359375, 0.17919921875, -0.096923828125, 0.267578125, 0.033935546875, 0.09429931640625, 0.492919921875, 0.453857421875, 0.38623046875, -0.0118408203125, 0.25634765625, 0.6826171875, -0.27490234375, 0.378662109375, 0.082275390625, 0.13720703125, -0.226806640625, 0.0577392578125, 0.316650390625, 0.28466796875, -0.7294921875, 0.5966796875, 0.1114501953125, -0.0882568359375, -0.01953125, -0.51513671875, -0.0028076171875, -0.0186767578125, 0.285888671875, -0.345947265625, -0.01483154296875, -0.185302734375, -0.18310546875, 0.4541015625, -0.048095703125, -0.302978515625, -0.60986328125, -0.405517578125, 0.28466796875, 0.094482421875, -0.1793212890625, 0.35400390625, -0.13916015625, 0.318115234375, -0.4150390625, -0.29443359375, 0.1776123046875, -0.28857421875, -0.4580078125, -0.6669921875, -0.6201171875, -0.0364990234375, 0.78564453125, 0.677734375, -0.25341796875, 0.0362548828125, 0.6640625, -0.689453125, 0.273193359375, -0.72705078125, -0.587890625, -0.5087890625, 0.0985107421875, 0.046142578125, -0.3837890625, -0.3330078125, 0.18212890625, -0.16552734375, 0.1533203125, 0.44677734375, 0.036773681640625, 0.1875, 0.012115478515625, -0.5068359375, 0.0791015625, -0.0572509765625, -0.0858154296875, 0.841796875, 0.046630859375, -0.27294921875, 0.50048828125, -0.6669921875, 0.475341796875, -0.26318359375, 0.58203125, -0.1365966796875, 0.05908203125, 0.1270751953125, -0.998046875, -0.1201171875, 0.0257568359375, 0.1787109375, -0.29541015625, -0.396240234375, 0.03338623046875, 1.138671875, 0.4140625, -0.64990234375, -0.2373046875, 0.736328125, -0.1612548828125, -0.17333984375, -0.0869140625, -0.414306640625, 0.2196044921875, -0.19677734375, -0.4375, 0.487060546875, 0.41064453125, -0.367919921875, -0.52099609375, 0.09844970703125, 0.1033935546875, -0.10687255859375, 0.14501953125, -0.301025390625, -0.064208984375, 0.19775390625, 0.177490234375, 0.35986328125, 0.41015625, 0.255126953125, -0.115966796875, 0.1640625, -0.02154541015625, 0.315185546875, 0.353515625, 0.476806640625, -0.263671875, -0.1510009765625, 0.1630859375, -0.26806640625, 0.001953125, -0.2325439453125, 0.343994140625, 0.421142578125, -0.29541015625, 0.427734375, -2.1171875, 0.55859375, 0.46484375, 0.1396484375, -0.5576171875, -0.1380615234375, 0.081787109375, -0.16357421875, 0.132080078125, 0.205078125, 0.41943359375, -0.164794921875, 0.82275390625, -0.345703125, 0.013671875, 0.61669921875, 0.252685546875, 0.27197265625, -0.18701171875, -0.40087890625, 0.195556640625, 0.4794921875, 0.180419921875, -0.0345458984375, -0.216552734375, 0.283935546875, 0.162353515625, 0.97509765625, 0.250732421875, 0.2998046875, -0.138916015625, 0.32763671875, 0.11962890625, 0.061767578125, 0.64208984375, -0.744140625, -0.3232421875, 0.5146484375, -0.59375, -0.00146484375, 0.8134765625, -0.422119140625, -0.2384033203125, -0.1162109375, -0.44091796875, -0.18359375, 0.1568603515625, -0.9052734375, -0.04034423828125, 0.0999755859375, 0.0115966796875, 0.58984375, -0.2802734375, 0.393798828125, -0.3212890625, -0.2091064453125, -0.1854248046875, -0.119384765625, 0.027587890625, -0.201171875, 0.162353515625, 0.10540771484375, 0.1033935546875, 0.27001953125, -0.216796875, -0.155029296875, -0.37841796875, 0.01549530029296875, 0.48681640625, 1.24609375, -0.162353515625, -0.3154296875, -0.5576171875, 0.55029296875, 0.33154296875, 0.0877685546875, 0.30517578125, -0.234619140625, -1.283203125, -0.0955810546875, 0.025146484375, -0.61181640625, 0.3623046875, 0.5458984375, -0.06396484375, 0.14501953125, -0.09063720703125, 0.111328125, 0.16357421875, 0.03094482421875, -0.51806640625, 0.298583984375, -0.138427734375, -0.427978515625, 0.1900634765625, -0.1199951171875, -0.31298828125, -0.23876953125, -0.213623046875, -0.1632080078125, -0.47265625, 0.065185546875, -0.0269775390625, -0.4912109375, -0.5205078125, -0.188720703125, -0.49365234375, -0.08795166015625, 0.2164306640625, -0.0966796875, -0.36181640625, -0.240478515625, -0.146240234375, -0.240478515625, 0.242919921875, 0.47509765625, 0.1611328125, 0.19970703125, -0.759765625, -0.33642578125, -0.53759765625, -0.2171630859375, 0.03619384765625, 0.201171875, 0.16748046875, 0.09637451171875, 0.424560546875, -0.0260162353515625, 0.1346435546875, -0.038818359375, -0.18798828125, 0.382080078125, 0.28515625, 0.6083984375, -0.1796875, 0.0419921875, -0.765625, -0.121337890625, 0.0311279296875, 0.19775390625, -0.2186279296875, -0.109130859375, -0.5380859375, 0.271484375, -1.4296875, 0.14990234375, -0.11065673828125, 1.2568359375, -0.1817626953125, 0.40478515625, 0.36669921875, 0.43212890625, -0.13427734375, -0.033416748046875, 0.25732421875, -0.0487060546875, 0.3544921875, 0.5888671875, -0.65234375, -0.2880859375, -0.1837158203125, -0.1964111328125, -0.105224609375, 0.1221923828125, 0.46484375, 0.022796630859375, -0.5869140625, 0.1602783203125, -0.1845703125, -0.0421142578125, -0.8076171875, 0.8583984375, -0.01025390625, -0.0418701171875, -0.01708984375, -0.75146484375, 0.447998046875, 0.18603515625, 0.5126953125, -0.010986328125, -0.65380859375, 0.6435546875, 0.30810546875, -0.12408447265625, 1.1123046875, 0.055419921875, -0.108154296875, -0.689453125, -0.2135009765625, -0.243896484375, 0.00244140625, 0.111572265625, 0.0853271484375, 0.431640625, -0.1591796875, 0.12451171875, 0.02001953125, -0.495849609375, -0.36474609375, 0.287841796875, -0.408203125, -0.26220703125, 0.110595703125, -0.40185546875, -0.164306640625, -0.029815673828125, -0.05731201171875, 0.9189453125, 0.0291748046875, -0.2646484375, -0.126953125, 0.193115234375, 0.1312255859375, -0.0516357421875, 0.294921875, -0.1412353515625, 0.23046875, 0.58349609375, -0.47607421875, 0.20263671875, -0.31591796875, -0.875, 0.200927734375, 0.2381591796875, 0.064697265625, -0.00421142578125, 0.055419921875, 0.1474609375, 0.1190185546875, -0.221923828125, 0.447021484375, 0.72021484375]}, {"frame_number": 10, "vector": [-0.12646484375, 0.29248046875, -0.014404296875, -0.1649169921875, -0.6259765625, 0.33642578125, 0.01123046875, -0.46484375, -0.14111328125, -0.03936767578125, -0.12744140625, 0.016357421875, 0.3525390625, 0.190185546875, -0.01220703125, 0.299072265625, 0.137939453125, -0.133544921875, -0.07794189453125, -0.44873046875, -0.5732421875, -6.1640625, -0.3193359375, -0.7138671875, 0.02130126953125, -0.0982666015625, -0.16064453125, -0.76171875, 0.469482421875, -0.17041015625, -0.9697265625, -0.6025390625, -0.7080078125, 0.001678466796875, -0.0836181640625, 0.1142578125, 0.318359375, -0.6826171875, -0.08697509765625, -0.461669921875, -0.07025146484375, 0.1903076171875, 0.22021484375, 0.048583984375, 0.9111328125, -0.35595703125, -0.42236328125, -0.22119140625, -0.5576171875, -0.124267578125, 0.06396484375, -0.28857421875, 0.0963134765625, 0.515625, 0.228515625, 0.02410888671875, 0.39990234375, -0.2470703125, 0.26123046875, -0.02056884765625, -0.06298828125, 0.2064208984375, 0.2666015625, 0.179931640625, -0.220947265625, -0.33984375, -0.496826171875, -0.0992431640625, 0.1829833984375, 0.1053466796875, 0.97705078125, -0.0946044921875, -0.04345703125, -0.0372314453125, -0.425537109375, 0.818359375, -0.372314453125, -0.45654296875, 0.17578125, -0.70068359375, 1.2001953125, 0.26513671875, 0.5615234375, 0.486328125, -0.265380859375, 0.038818359375, 0.2022705078125, 0.016357421875, 0.1416015625, -0.611328125, -0.208740234375, -0.0958251953125, 0.07666015625, -0.17236328125, 0.28076171875, 0.328125, 0.08837890625, 0.5439453125, 0.0177459716796875, -0.67041015625, 0.763671875, 0.33837890625, -0.3681640625, -0.6240234375, -0.00439453125, -0.144287109375, -0.00738525390625, -0.15234375, -0.142333984375, -0.047119140625, -0.0782470703125, 0.165771484375, 0.0841064453125, -0.52734375, 0.134765625, -0.5703125, -0.247802734375, -0.14892578125, 0.32275390625, 0.517578125, 0.0516357421875, -0.226806640625, -0.0390625, -0.70361328125, -0.0784912109375, 0.225341796875, -0.122802734375, 0.379638671875, 0.20751953125, 0.36181640625, -0.03955078125, -0.408203125, 0.04193115234375, 0.08984375, 0.5419921875, -0.0322265625, 0.085693359375, 0.77490234375, 0.580078125, 0.02685546875, 0.064453125, 0.294189453125, 0.1611328125, -0.09326171875, -0.248046875, 0.1319580078125, -0.0095672607421875, -0.33984375, 0.230224609375, 1.0283203125, -0.14404296875, -1.552734375, 0.0736083984375, -0.41845703125, 0.427490234375, -0.267333984375, -0.267578125, 0.93798828125, 0.296875, -0.43310546875, -0.259765625, 0.30322265625, 0.470947265625, 0.3193359375, -0.3056640625, -0.31103515625, -0.0205078125, 0.689453125, 0.078857421875, 0.0997314453125, 0.1729736328125, 0.4609375, -0.1795654296875, 0.62255859375, -0.258056640625, 1.0283203125, -1.09375, 0.110107421875, -0.0162353515625, -0.086669921875, 0.0478515625, 0.2158203125, -0.1551513671875, 0.038330078125, 0.068115234375, -0.0106201171875, 0.017822265625, 0.1927490234375, -0.107421875, 0.2198486328125, 0.162841796875, 0.1536865234375, 0.458251953125, -0.3564453125, 0.5322265625, -0.2061767578125, 0.421875, -0.30322265625, -0.365234375, 1.099609375, 0.461669921875, -0.01416015625, -0.2294921875, 0.10205078125, -0.048095703125, 0.09088134765625, 0.65625, 0.020782470703125, -0.6435546875, -0.111572265625, 0.2266845703125, -0.112548828125, 0.255615234375, -0.8173828125, -0.52001953125, -0.494140625, 0.49267578125, 1.005859375, -0.00841522216796875, -0.402099609375, -0.4130859375, -0.5478515625, -0.136962890625, -0.61669921875, -0.173095703125, 0.17578125, -0.1290283203125, -0.22900390625, 0.77294921875, 0.02276611328125, -0.0994873046875, 0.27734375, 0.062255859375, 0.282470703125, -0.28564453125, -0.525390625, 0.78564453125, -0.0892333984375, -0.92724609375, -0.07958984375, 0.705078125, 0.2254638671875, 0.0606689453125, 0.154052734375, -0.16796875, 0.9970703125, -0.18310546875, 0.0377197265625, -0.40234375, -0.09521484375, -0.026123046875, -0.04986572265625, 0.328125, -0.072509765625, -0.712890625, -0.111328125, -0.176025390625, 0.0032958984375, 0.437744140625, 0.4697265625, -0.05877685546875, -0.1845703125, -0.05908203125, -0.0400390625, -0.3232421875, 0.0400390625, -0.78173828125, -0.354248046875, 0.55859375, -0.45361328125, 0.03515625, 0.433837890625, 0.2861328125, 0.225830078125, -0.2342529296875, -0.68896484375, -0.07586669921875, 0.149169921875, -0.373046875, -0.394287109375, -0.31787109375, 0.6494140625, -0.44287109375, 0.143798828125, 0.384521484375, 0.302978515625, -0.12005615234375, 0.55908203125, -0.58740234375, 1.1474609375, -0.01800537109375, 0.0599365234375, 0.0438232421875, -0.388671875, 0.09649658203125, 0.1717529296875, 0.1943359375, 0.35302734375, 0.66015625, -0.26220703125, 0.0229644775390625, 0.7421875, -0.0302734375, -0.04052734375, -0.31201171875, 0.3671875, 0.34619140625, -0.66796875, 0.13525390625, -0.098876953125, 0.346923828125, -0.32470703125, 0.000732421875, -0.31005859375, -0.364013671875, 0.3828125, -0.082275390625, -0.00750732421875, -0.28857421875, 0.206787109375, -0.93212890625, 0.0899658203125, 0.0987548828125, -0.455078125, -0.88427734375, 0.6455078125, -0.45556640625, 0.328125, -0.61962890625, -0.040771484375, -0.477783203125, -0.064453125, -0.208984375, -0.48291015625, 0.2449951171875, 0.2313232421875, 0.10736083984375, -0.60986328125, 0.00982666015625, -0.00244140625, -0.71875, 0.01708984375, -0.296630859375, -0.6650390625, -0.5029296875, 0.2265625, -0.46142578125, 0.3662109375, 0.1953125, 0.056396484375, -0.36181640625, -0.289794921875, 0.253173828125, 0.152587890625, 0.445556640625, 0.3369140625, 0.240966796875, -0.1627197265625, -0.3271484375, -0.237548828125, -0.1566162109375, 0.045654296875, 0.48876953125, 0.28125, -0.6123046875, -0.452392578125, -0.049560546875, 0.2081298828125, 0.07421875, -0.0831298828125, -0.026611328125, 0.26318359375, -0.1123046875, 0.3369140625, 0.287841796875, -0.0576171875, 0.394775390625, -0.2490234375, 0.1494140625, -0.5390625, -0.0130615234375, -0.41650390625, 0.1162109375, 0.74267578125, 0.214599609375, -0.0499267578125, -0.02105712890625, -0.14794921875, -0.2822265625, 0.537109375, -0.634765625, -0.228271484375, 0.27587890625, 0.251953125, 1.0634765625, 0.64794921875, -0.0615234375, -0.65234375, -0.0433349609375, -0.3720703125, 0.219970703125, 0.1455078125, -0.119140625, -0.44189453125, -0.217041015625, 0.00830078125, -0.3759765625, 0.02008056640625, 0.36865234375, -0.0440673828125, 0.175537109375, -0.892578125, -0.7333984375, -0.2978515625, -0.38525390625, 0.5927734375, 0.7314453125, -0.303955078125, 1.0322265625, -0.31591796875, 0.134033203125, -0.03228759765625, -0.245361328125, -0.084228515625, -0.0198974609375, -0.486328125, -1.404296875, -0.08880615234375, 0.0966796875, 0.1771240234375, 0.07568359375, -6.1640625, -0.44287109375, -0.058349609375, -0.7978515625, -0.2109375, -0.114990234375, -1.23046875, -0.63671875, -0.253173828125, -0.013671875, 0.59326171875, 0.6591796875, -0.27490234375, -0.004119873046875, 0.18408203125, 0.214599609375, 0.0469970703125, -0.02294921875, 0.18896484375, 0.011474609375, 0.193359375, -0.12432861328125, 0.090576171875, 0.402099609375, -0.28955078125, -0.2105712890625, 0.39306640625, -0.6650390625, -0.60302734375, -0.08642578125, 0.09228515625, -0.28857421875, -0.36865234375, -0.4443359375, -0.03662109375, 0.60302734375, 0.796875, -0.37255859375, -0.4072265625, 0.21826171875, -1.0078125, 0.31640625, -0.03857421875, 0.0159912109375, 0.2138671875, 0.319580078125, -0.061431884765625, -0.751953125, -0.75634765625, 0.129150390625, 0.0867919921875, -1.0478515625, -0.072509765625, 0.349365234375, -0.0096435546875, 0.24755859375, -0.056854248046875, 0.10205078125, 0.1734619140625, 0.18505859375, 0.54296875, 0.337646484375, 0.435546875, -0.466064453125, 0.0, 0.0372314453125, -0.0001220703125, 0.088623046875, 0.0697021484375, -0.094482421875, 0.027587890625, 0.65185546875, -0.1259765625, -0.24365234375, -0.15283203125, -0.044677734375, 0.1153564453125, -0.17431640625, 0.061767578125, 0.208984375, 0.0235595703125, 0.167724609375, 0.41162109375, -0.21875, -0.1650390625, -0.13818359375, -0.345947265625, -0.626953125, -0.1292724609375, 0.7275390625, 0.224609375, -0.208740234375, 0.060546875, 0.2529296875, 1.2666015625, 0.189208984375, 0.1353759765625, 0.1226806640625, 0.13134765625, -0.322998046875, -0.497314453125, 0.01513671875, -0.065185546875, 0.35693359375, 0.265625, 0.07373046875, -0.0272216796875, 0.3505859375, -0.04833984375, 0.74951171875, 0.12939453125, 0.35400390625, 0.490966796875, -0.2266845703125, 0.486328125, -1.5576171875, -0.38134765625, -0.2080078125, 0.1036376953125, -0.0072021484375, -0.3212890625, 0.72119140625, 0.2607421875, -0.1112060546875, 0.088623046875, 0.320068359375, -0.060302734375, -0.242431640625, -0.9482421875, -0.45263671875, -0.287353515625, -0.43408203125, -0.9521484375, 0.29638671875, 0.32080078125, 0.51953125, -0.3779296875, -0.096435546875, 0.05908203125, 0.0057373046875, 0.366943359375, -0.23828125, 0.11029052734375, 0.1748046875, 0.355224609375, -0.36572265625, 0.1953125, -0.29833984375, -0.447998046875, -0.2763671875, 0.1060791015625, -0.119384765625, -0.1033935546875, -0.00396728515625, -0.114501953125, 0.00537109375, 0.423583984375, -0.3330078125, 0.230712890625, 0.40283203125, 0.051513671875, -0.2481689453125, 0.01470947265625, 0.0811767578125, 0.06298828125, -0.0810546875, -0.204345703125, -1.490234375, -0.45849609375, 0.2041015625, -0.07421875, -0.30615234375, -0.37353515625, -0.19189453125, -0.11572265625, -0.138427734375, -0.27392578125, -0.22314453125, -0.8759765625, 0.2396240234375, 0.39501953125, -0.2222900390625, -0.1806640625, -0.1070556640625, -0.583984375, 0.23486328125, 0.26416015625, -0.07012939453125, -0.2646484375, 0.03125, -0.0693359375, 0.04962158203125, 0.05828857421875, -0.224853515625, -0.04473876953125, 0.046173095703125, 0.33740234375, 0.2724609375, -0.097412109375, 0.1214599609375, 0.014617919921875, -0.371826171875, 0.6484375, 0.0950927734375, -0.14208984375, -0.30126953125, 1.3369140625, 0.095703125, -0.2763671875, 0.026123046875, -0.51416015625, 0.3798828125, -0.2435302734375, -0.21728515625, 0.330078125, 0.152099609375, -0.0168609619140625, 0.27490234375, 0.127685546875, 0.251220703125, -0.343017578125, 0.36474609375, 0.24072265625, -0.671875, 0.20361328125, 0.310791015625, -0.09765625, -0.26806640625, -0.26025390625, -0.29296875, -0.841796875, 0.003173828125, -0.117919921875, -0.1058349609375, -0.288330078125, 0.06982421875, -0.076171875, 0.480224609375, -0.455078125, -1.419921875, -0.0914306640625, -0.1129150390625, 0.22314453125, 0.27587890625, -0.07989501953125, 0.3388671875, 0.1593017578125, -0.163330078125, 0.279541015625, -0.01507568359375, 0.168701171875, 0.44482421875, 0.00830078125, 0.143310546875, -0.229736328125, 0.3095703125, 0.2110595703125, -0.405029296875, -0.6201171875, -0.206787109375, -0.46240234375, 0.04248046875, 0.25048828125, -0.272705078125, 0.81201171875, -0.06884765625, 0.0999755859375, -0.884765625, 0.00592041015625, 0.027587890625, 0.1884765625, -0.6552734375, 0.308349609375, -0.53271484375, 0.38525390625, -0.0341796875, 0.16650390625, -0.1624755859375, -0.50390625, -0.681640625, 0.1239013671875, -0.456787109375, 0.448974609375, -0.3544921875, 0.935546875, -0.32958984375, 0.0672607421875, -0.11669921875, -0.0587158203125, -0.43115234375, 0.135009765625, 1.4140625, 0.021697998046875, 0.26318359375, -0.1246337890625, 0.40625, -1.080078125, -0.2476806640625, 0.54296875, 0.05023193359375, -0.19482421875, 0.568359375, 0.16943359375, 0.0711669921875, 0.2021484375, 0.044677734375, 0.18017578125, -0.05322265625, 0.48876953125, -0.5361328125, 0.032470703125, -0.62744140625, 0.0548095703125, 0.078369140625, -0.45263671875, 0.0091552734375, -0.057861328125, -0.07275390625, 0.0699462890625, -0.8994140625, -0.2064208984375, 0.24365234375, -0.11810302734375, 0.265625, -0.0079345703125, -0.111572265625, -0.84765625, 0.0528564453125, 0.329345703125, -0.01885986328125, 0.03155517578125, -0.2247314453125, -0.5087890625, 0.60791015625, -0.409423828125, 0.0177001953125, -0.64306640625, 0.00067138671875, -0.38671875, 0.12078857421875, -0.03125, 0.086181640625, 0.0838623046875, 0.19873046875, -2.28125, 0.342041015625, -0.190673828125, 0.268310546875, 0.68017578125, -0.0025634765625, -0.31689453125, -0.2314453125, 0.19091796875, -0.0980224609375, 0.16455078125, 0.03564453125, 0.031982421875, 0.48095703125, 0.461181640625, 0.30419921875, 0.0142822265625, 0.28125, 0.66259765625, -0.2607421875, 0.344482421875, 0.0888671875, 0.0985107421875, -0.3154296875, 0.137939453125, 0.267578125, 0.2294921875, -0.7333984375, 0.6064453125, 0.169677734375, -0.1317138671875, 0.007568359375, -0.4658203125, 0.003173828125, -0.0335693359375, 0.3076171875, -0.341552734375, 0.0003662109375, -0.212158203125, -0.23583984375, 0.408935546875, -0.10321044921875, -0.3125, -0.6279296875, -0.392822265625, 0.25, 0.1126708984375, -0.1669921875, 0.393310546875, -0.1309814453125, 0.3095703125, -0.410888671875, -0.285888671875, 0.1749267578125, -0.29736328125, -0.51318359375, -0.6474609375, -0.64111328125, -0.01934814453125, 0.79248046875, 0.5927734375, -0.243896484375, -0.0096435546875, 0.65576171875, -0.6767578125, 0.2412109375, -0.66064453125, -0.64306640625, -0.50634765625, 0.11474609375, 0.08868408203125, -0.31640625, -0.28662109375, 0.1810302734375, -0.13818359375, 0.16259765625, 0.430419921875, 0.03466796875, 0.1630859375, -0.00344085693359375, -0.4833984375, 0.08154296875, -0.0653076171875, -0.0841064453125, 0.7763671875, 0.065185546875, -0.275634765625, 0.52099609375, -0.64111328125, 0.49462890625, -0.287109375, 0.5908203125, -0.132568359375, 0.10009765625, 0.104736328125, -0.986328125, -0.128662109375, 0.05908203125, 0.160400390625, -0.28076171875, -0.4658203125, 0.036865234375, 1.13671875, 0.4150390625, -0.638671875, -0.256103515625, 0.7734375, -0.187255859375, -0.17724609375, -0.126220703125, -0.396484375, 0.236328125, -0.174560546875, -0.423583984375, 0.441162109375, 0.41357421875, -0.353271484375, -0.5400390625, 0.0889892578125, 0.0614013671875, -0.156982421875, 0.1533203125, -0.1978759765625, -0.0810546875, 0.2213134765625, 0.265625, 0.35986328125, 0.34130859375, 0.23095703125, -0.124267578125, 0.164794921875, -0.00958251953125, 0.345703125, 0.339599609375, 0.51611328125, -0.295166015625, -0.11376953125, 0.129638671875, -0.316162109375, -0.013671875, -0.202880859375, 0.38720703125, 0.370361328125, -0.29248046875, 0.442626953125, -2.12890625, 0.541015625, 0.469970703125, 0.166259765625, -0.5986328125, -0.12451171875, 0.072998046875, -0.166015625, 0.1512451171875, 0.24560546875, 0.3681640625, -0.09423828125, 0.81396484375, -0.352294921875, 0.093994140625, 0.6669921875, 0.262451171875, 0.2890625, -0.21630859375, -0.37255859375, 0.1766357421875, 0.541015625, 0.172119140625, -0.0679931640625, -0.185546875, 0.253173828125, 0.131591796875, 0.99755859375, 0.30126953125, 0.27734375, -0.2646484375, 0.32666015625, 0.13818359375, 0.01123046875, 0.6650390625, -0.7392578125, -0.31884765625, 0.509765625, -0.61669921875, 0.05926513671875, 0.7333984375, -0.4765625, -0.220703125, -0.0618896484375, -0.4228515625, -0.232421875, 0.1492919921875, -0.904296875, -0.06005859375, 0.15771484375, 0.028076171875, 0.60546875, -0.2119140625, 0.37060546875, -0.3212890625, -0.2369384765625, -0.193115234375, -0.095458984375, 0.037353515625, -0.16162109375, 0.1285400390625, 0.0894775390625, 0.15185546875, 0.34814453125, -0.15771484375, -0.134033203125, -0.341552734375, 0.026336669921875, 0.462890625, 1.240234375, -0.149658203125, -0.264892578125, -0.6376953125, 0.58984375, 0.38037109375, 0.0750732421875, 0.41064453125, -0.237548828125, -1.3642578125, -0.10064697265625, 0.048583984375, -0.609375, 0.393798828125, 0.5732421875, -0.06097412109375, 0.1728515625, -0.1214599609375, 0.0552978515625, 0.138916015625, -0.00177001953125, -0.52294921875, 0.2509765625, -0.1485595703125, -0.461181640625, 0.171142578125, -0.144287109375, -0.328125, -0.2042236328125, -0.18603515625, -0.236328125, -0.470947265625, 0.05322265625, 0.0050048828125, -0.43603515625, -0.5166015625, -0.1279296875, -0.4931640625, -0.0606689453125, 0.20361328125, -0.087158203125, -0.34765625, -0.213623046875, -0.158447265625, -0.2578125, 0.2158203125, 0.5048828125, 0.108154296875, 0.2119140625, -0.7421875, -0.338623046875, -0.53271484375, -0.23046875, 0.0333251953125, 0.164306640625, 0.17724609375, 0.0782470703125, 0.42626953125, -0.06622314453125, 0.1334228515625, 0.00408935546875, -0.1630859375, 0.45947265625, 0.218017578125, 0.60546875, -0.1436767578125, 0.041015625, -0.7236328125, -0.129150390625, 0.0418701171875, 0.16845703125, -0.21142578125, -0.10528564453125, -0.49267578125, 0.303955078125, -1.478515625, 0.0955810546875, -0.1572265625, 1.23828125, -0.168701171875, 0.412353515625, 0.376953125, 0.45751953125, -0.1651611328125, -0.077880859375, 0.270263671875, -0.0435791015625, 0.4248046875, 0.626953125, -0.623046875, -0.29150390625, -0.135986328125, -0.1832275390625, -0.044097900390625, 0.187744140625, 0.462890625, 0.01690673828125, -0.576171875, 0.161376953125, -0.1787109375, -0.05950927734375, -0.779296875, 0.89013671875, -0.0537109375, -0.041259765625, 0.0096435546875, -0.7001953125, 0.474609375, 0.1929931640625, 0.46435546875, -0.055419921875, -0.68994140625, 0.7421875, 0.234375, -0.10302734375, 1.0888671875, 0.0706787109375, -0.0791015625, -0.783203125, -0.21630859375, -0.281982421875, -0.01318359375, 0.135009765625, 0.081298828125, 0.4365234375, -0.17529296875, 0.1168212890625, -0.01824951171875, -0.488037109375, -0.352294921875, 0.345703125, -0.4189453125, -0.3359375, 0.10577392578125, -0.41650390625, -0.137939453125, -0.0972900390625, -0.05780029296875, 0.86181640625, -0.0137939453125, -0.2724609375, -0.0938720703125, 0.1708984375, 0.06951904296875, -0.024169921875, 0.27001953125, -0.1231689453125, 0.2469482421875, 0.603515625, -0.4736328125, 0.207275390625, -0.293701171875, -0.8330078125, 0.2147216796875, 0.290771484375, 0.086669921875, -0.04351806640625, 0.0748291015625, 0.14990234375, 0.12158203125, -0.216064453125, 0.480224609375, 0.634765625]}, {"frame_number": 11, "vector": [-0.12127685546875, 0.3115234375, -0.0205078125, -0.152099609375, -0.61328125, 0.317626953125, -0.020721435546875, -0.47705078125, -0.107421875, 0.02105712890625, -0.136474609375, 0.0908203125, 0.32373046875, 0.1826171875, -0.031005859375, 0.314453125, 0.167236328125, -0.16259765625, -0.037689208984375, -0.47998046875, -0.54296875, -6.078125, -0.349609375, -0.77490234375, 0.081787109375, -0.0548095703125, -0.15966796875, -0.77099609375, 0.50390625, -0.20458984375, -0.94091796875, -0.623046875, -0.6865234375, -0.03692626953125, -0.1058349609375, 0.071044921875, 0.31884765625, -0.63037109375, -0.07012939453125, -0.393310546875, -0.02703857421875, 0.20751953125, 0.2568359375, 0.040771484375, 0.90283203125, -0.41259765625, -0.430419921875, -0.20751953125, -0.5869140625, -0.1640625, 0.0482177734375, -0.263427734375, 0.1800537109375, 0.55712890625, 0.222900390625, -0.01123046875, 0.438720703125, -0.2010498046875, 0.248779296875, -0.02520751953125, -0.087646484375, 0.202392578125, 0.298828125, 0.19091796875, -0.2021484375, -0.36376953125, -0.487060546875, -0.0714111328125, 0.1710205078125, 0.0252685546875, 0.93017578125, -0.0560302734375, -0.01318359375, -0.097412109375, -0.461669921875, 0.8369140625, -0.299072265625, -0.4853515625, 0.25244140625, -0.68603515625, 1.251953125, 0.271728515625, 0.572265625, 0.50634765625, -0.258056640625, 0.0406494140625, 0.135498046875, -0.051177978515625, 0.16650390625, -0.578125, -0.1881103515625, -0.15380859375, 0.06396484375, -0.1455078125, 0.281494140625, 0.37353515625, 0.086181640625, 0.59033203125, 0.030731201171875, -0.64306640625, 0.7578125, 0.383056640625, -0.3974609375, -0.63330078125, 0.00653076171875, -0.0965576171875, -0.03204345703125, -0.12890625, -0.18359375, -0.046630859375, -0.0247802734375, 0.1806640625, 0.1083984375, -0.53125, 0.1329345703125, -0.6240234375, -0.259521484375, -0.164306640625, 0.35498046875, 0.50439453125, 0.0433349609375, -0.27001953125, -0.0135498046875, -0.6923828125, -0.0867919921875, 0.255859375, -0.1209716796875, 0.35693359375, 0.167236328125, 0.29052734375, -0.064453125, -0.37548828125, 0.069580078125, 0.05523681640625, 0.60205078125, -0.0093994140625, 0.08465576171875, 0.75927734375, 0.525390625, -0.0244140625, 0.08251953125, 0.291748046875, 0.17138671875, -0.096923828125, -0.237548828125, 0.0897216796875, -0.04046630859375, -0.3388671875, 0.1876220703125, 1.04296875, -0.1998291015625, -1.5419921875, 0.0509033203125, -0.33251953125, 0.39697265625, -0.2236328125, -0.24462890625, 0.9140625, 0.2122802734375, -0.41748046875, -0.265869140625, 0.299560546875, 0.439453125, 0.312255859375, -0.35546875, -0.2890625, 0.0081787109375, 0.6220703125, 0.0499267578125, 0.095947265625, 0.1605224609375, 0.4169921875, -0.174560546875, 0.63330078125, -0.25, 1.0625, -1.126953125, 0.220458984375, -0.0419921875, -0.108154296875, 0.005615234375, 0.169677734375, -0.1944580078125, 0.010467529296875, 0.078369140625, 0.01873779296875, 0.045166015625, 0.208984375, -0.091552734375, 0.1895751953125, 0.1910400390625, 0.107421875, 0.442626953125, -0.38720703125, 0.5556640625, -0.138671875, 0.421142578125, -0.17724609375, -0.2705078125, 1.203125, 0.44580078125, -0.0400390625, -0.1944580078125, 0.0894775390625, 0.02655029296875, 0.02410888671875, 0.623046875, 0.00042724609375, -0.69384765625, -0.1231689453125, 0.22265625, -0.058349609375, 0.2216796875, -0.85595703125, -0.48486328125, -0.529296875, 0.464111328125, 1.025390625, -0.004852294921875, -0.4599609375, -0.407470703125, -0.54736328125, -0.1026611328125, -0.564453125, -0.21044921875, 0.26025390625, -0.150390625, -0.28662109375, 0.78466796875, 0.0179443359375, -0.158935546875, 0.21142578125, 0.086669921875, 0.19189453125, -0.394287109375, -0.51953125, 0.76025390625, -0.0513916015625, -0.99755859375, -0.055419921875, 0.70556640625, 0.24951171875, -0.0496826171875, 0.111572265625, -0.182861328125, 0.9306640625, -0.237060546875, 0.0074462890625, -0.36181640625, -0.146240234375, -0.066650390625, -0.03912353515625, 0.26708984375, -0.05517578125, -0.65478515625, -0.099609375, -0.177001953125, -0.050048828125, 0.486328125, 0.454833984375, 0.009521484375, -0.225341796875, -0.08740234375, 0.0157470703125, -0.234375, 0.0272216796875, -0.77001953125, -0.3134765625, 0.615234375, -0.462890625, 0.0283203125, 0.391845703125, 0.29443359375, 0.30419921875, -0.25146484375, -0.72119140625, -0.1026611328125, 0.1380615234375, -0.351318359375, -0.364501953125, -0.310546875, 0.65234375, -0.453857421875, 0.09423828125, 0.361572265625, 0.30126953125, -0.1317138671875, 0.50390625, -0.5146484375, 1.166015625, 0.038818359375, 0.04241943359375, 0.0888671875, -0.385009765625, 0.05712890625, 0.1314697265625, 0.18701171875, 0.389404296875, 0.611328125, -0.24267578125, -0.02386474609375, 0.720703125, 0.03704833984375, -0.004150390625, -0.28466796875, 0.362548828125, 0.3310546875, -0.7177734375, 0.1011962890625, -0.162841796875, 0.365478515625, -0.379638671875, 0.015380859375, -0.341552734375, -0.334716796875, 0.3857421875, -0.0673828125, 0.00225830078125, -0.30810546875, 0.2470703125, -0.927734375, 0.10205078125, 0.07073974609375, -0.453369140625, -0.861328125, 0.62109375, -0.4775390625, 0.36279296875, -0.56982421875, 0.00927734375, -0.48974609375, -0.07763671875, -0.1942138671875, -0.46875, 0.1522216796875, 0.20947265625, 0.15869140625, -0.615234375, 0.02874755859375, -0.0130615234375, -0.71630859375, 0.034912109375, -0.3359375, -0.7255859375, -0.48681640625, 0.161865234375, -0.49658203125, 0.27783203125, 0.20654296875, 0.0325927734375, -0.3017578125, -0.39453125, 0.196044921875, 0.16552734375, 0.414306640625, 0.337890625, 0.23974609375, -0.1627197265625, -0.352783203125, -0.21630859375, -0.1767578125, -0.0064697265625, 0.525390625, 0.3115234375, -0.63916015625, -0.4443359375, -0.087890625, 0.25732421875, 0.063720703125, -0.1180419921875, -0.0115966796875, 0.2685546875, -0.119873046875, 0.3310546875, 0.297119140625, -0.0313720703125, 0.39306640625, -0.2105712890625, 0.2509765625, -0.5146484375, -0.032958984375, -0.37939453125, 0.1492919921875, 0.6416015625, 0.1658935546875, -0.115966796875, 0.0426025390625, -0.189697265625, -0.26318359375, 0.498779296875, -0.6611328125, -0.2236328125, 0.3154296875, 0.2388916015625, 1.044921875, 0.6474609375, -0.077392578125, -0.62451171875, -0.12451171875, -0.44287109375, 0.20654296875, 0.09686279296875, -0.134033203125, -0.46337890625, -0.141845703125, 0.01025390625, -0.311767578125, 0.07281494140625, 0.36181640625, 0.0120849609375, 0.21875, -0.91650390625, -0.697265625, -0.2685546875, -0.406494140625, 0.5673828125, 0.7373046875, -0.29541015625, 1.1162109375, -0.2232666015625, 0.1727294921875, 0.00286865234375, -0.156005859375, -0.104248046875, -0.08544921875, -0.513671875, -1.41015625, -0.105712890625, 0.1033935546875, 0.18798828125, 0.1087646484375, -6.0859375, -0.44091796875, -0.03076171875, -0.7890625, -0.155029296875, -0.0908203125, -1.205078125, -0.654296875, -0.240966796875, 0.0201416015625, 0.5615234375, 0.650390625, -0.200927734375, 0.02276611328125, 0.1376953125, 0.240966796875, 0.0240478515625, -0.033447265625, 0.239501953125, 0.08203125, 0.205810546875, -0.1236572265625, 0.0946044921875, 0.43212890625, -0.339599609375, -0.216796875, 0.362548828125, -0.68359375, -0.66015625, -0.08837890625, 0.060302734375, -0.328857421875, -0.3408203125, -0.416015625, -0.032470703125, 0.5908203125, 0.77197265625, -0.32421875, -0.39111328125, 0.212890625, -0.98388671875, 0.33642578125, -0.07568359375, -0.000518798828125, 0.170166015625, 0.387939453125, -0.09051513671875, -0.7734375, -0.72021484375, 0.139892578125, 0.1650390625, -1.0205078125, -0.034912109375, 0.34130859375, 0.00384521484375, 0.284912109375, -0.0531005859375, 0.1488037109375, 0.16162109375, 0.147216796875, 0.489990234375, 0.32080078125, 0.4521484375, -0.3564453125, 0.01806640625, 0.054840087890625, -0.0260009765625, 0.12451171875, 0.0556640625, -0.0577392578125, 0.0523681640625, 0.71630859375, -0.1796875, -0.23095703125, -0.129150390625, -0.0924072265625, 0.0721435546875, -0.2158203125, 0.077392578125, 0.242919921875, 0.0347900390625, 0.1116943359375, 0.3603515625, -0.235595703125, -0.1923828125, -0.067626953125, -0.357666015625, -0.62451171875, -0.110107421875, 0.716796875, 0.1884765625, -0.193603515625, 0.1192626953125, 0.244384765625, 1.296875, 0.185546875, 0.06689453125, 0.1522216796875, 0.0628662109375, -0.3583984375, -0.4853515625, -0.0625, -0.102294921875, 0.3134765625, 0.283203125, 0.100341796875, -0.02386474609375, 0.412109375, -0.06591796875, 0.79833984375, 0.153076171875, 0.387939453125, 0.54296875, -0.188720703125, 0.46142578125, -1.490234375, -0.447509765625, -0.213134765625, 0.09912109375, -0.0556640625, -0.336181640625, 0.77685546875, 0.3076171875, -0.0980224609375, 0.111328125, 0.26513671875, -0.07958984375, -0.174072265625, -0.931640625, -0.405029296875, -0.2421875, -0.4287109375, -0.96142578125, 0.283203125, 0.353515625, 0.458984375, -0.34716796875, -0.062744140625, 0.042236328125, -0.05078125, 0.337890625, -0.240234375, 0.10418701171875, 0.18701171875, 0.402587890625, -0.394287109375, 0.23095703125, -0.33349609375, -0.473388671875, -0.3369140625, 0.1983642578125, -0.165771484375, -0.1099853515625, 0.0914306640625, -0.2158203125, 0.01171875, 0.45947265625, -0.33447265625, 0.1920166015625, 0.33544921875, 0.00798797607421875, -0.256103515625, 0.03955078125, 0.109375, 0.00030517578125, -0.1044921875, -0.228271484375, -1.5078125, -0.45361328125, 0.161865234375, -0.012451171875, -0.35986328125, -0.34033203125, -0.2425537109375, -0.095947265625, -0.048492431640625, -0.251953125, -0.1650390625, -0.8564453125, 0.189208984375, 0.33837890625, -0.1834716796875, -0.1865234375, -0.1549072265625, -0.6181640625, 0.21533203125, 0.28564453125, -0.093017578125, -0.267578125, -0.00335693359375, -0.0537109375, 0.045257568359375, 0.075927734375, -0.25, -0.0124664306640625, 0.038604736328125, 0.459716796875, 0.24462890625, -0.0545654296875, 0.1744384765625, -0.00103759765625, -0.372802734375, 0.6982421875, 0.0511474609375, -0.167724609375, -0.267822265625, 1.3310546875, 0.049560546875, -0.323974609375, -0.0157470703125, -0.372314453125, 0.346435546875, -0.2064208984375, -0.1845703125, 0.365478515625, 0.19873046875, -0.060791015625, 0.25146484375, 0.1678466796875, 0.26611328125, -0.329345703125, 0.391357421875, 0.2314453125, -0.6650390625, 0.187744140625, 0.2978515625, -0.074462890625, -0.2548828125, -0.19873046875, -0.3056640625, -0.8515625, -0.0040283203125, -0.11993408203125, -0.1422119140625, -0.3193359375, 0.053466796875, -0.107177734375, 0.414306640625, -0.4755859375, -1.3603515625, -0.103515625, -0.146240234375, 0.233154296875, 0.263671875, -0.11529541015625, 0.316162109375, 0.20361328125, -0.17822265625, 0.300048828125, -0.0186767578125, 0.156982421875, 0.4443359375, 0.008544921875, 0.134521484375, -0.1484375, 0.314697265625, 0.2734375, -0.337890625, -0.6083984375, -0.177734375, -0.43310546875, 0.010498046875, 0.3017578125, -0.31005859375, 0.8154296875, -0.04931640625, 0.11090087890625, -0.77001953125, -0.0300750732421875, -0.0213623046875, 0.192138671875, -0.55859375, 0.3046875, -0.59375, 0.40087890625, -0.0341796875, 0.180908203125, -0.11688232421875, -0.564453125, -0.7744140625, 0.093994140625, -0.40234375, 0.4462890625, -0.330078125, 0.96875, -0.33154296875, 0.081787109375, -0.137451171875, -0.0289306640625, -0.427734375, 0.105224609375, 1.41015625, 0.041748046875, 0.254150390625, -0.00238037109375, 0.39501953125, -1.1455078125, -0.2607421875, 0.53125, 0.04833984375, -0.258056640625, 0.5498046875, 0.1856689453125, 0.104736328125, 0.166748046875, 0.0859375, 0.174560546875, -0.0394287109375, 0.4296875, -0.394775390625, 0.039306640625, -0.69873046875, 0.0784912109375, 0.098876953125, -0.49072265625, 0.00408935546875, -0.1055908203125, -0.138427734375, 0.113037109375, -0.853515625, -0.3076171875, 0.18896484375, -0.0750732421875, 0.184326171875, -0.05987548828125, -0.148193359375, -0.7705078125, 0.044677734375, 0.3076171875, -0.08819580078125, 0.0758056640625, -0.225341796875, -0.54541015625, 0.5859375, -0.43408203125, 0.01422119140625, -0.650390625, 0.0242919921875, -0.33740234375, 0.1063232421875, -0.00439453125, -0.020263671875, 0.093994140625, 0.2177734375, -2.259765625, 0.354736328125, -0.29541015625, 0.287353515625, 0.68310546875, -0.0780029296875, -0.3623046875, -0.21337890625, 0.1884765625, -0.089599609375, 0.22314453125, 0.101806640625, 0.10137939453125, 0.5302734375, 0.48974609375, 0.412841796875, -0.04443359375, 0.250732421875, 0.70263671875, -0.2880859375, 0.341796875, 0.060546875, 0.15234375, -0.21484375, 0.076171875, 0.31103515625, 0.3037109375, -0.74755859375, 0.615234375, 0.136474609375, -0.1519775390625, 0.01025390625, -0.53466796875, 0.015869140625, -0.00653076171875, 0.26611328125, -0.3388671875, -0.021636962890625, -0.200927734375, -0.21337890625, 0.5048828125, -0.00836181640625, -0.3046875, -0.6337890625, -0.40185546875, 0.30322265625, 0.1295166015625, -0.1768798828125, 0.38818359375, -0.14208984375, 0.255126953125, -0.39013671875, -0.264892578125, 0.172607421875, -0.296142578125, -0.5048828125, -0.65869140625, -0.630859375, -0.03521728515625, 0.79345703125, 0.60546875, -0.253662109375, 0.0316162109375, 0.6826171875, -0.6875, 0.306640625, -0.6806640625, -0.6005859375, -0.5224609375, 0.0484619140625, 0.0732421875, -0.386474609375, -0.338134765625, 0.2061767578125, -0.1796875, 0.1376953125, 0.429931640625, 0.055328369140625, 0.1854248046875, 0.00775146484375, -0.474609375, 0.087646484375, -0.01202392578125, -0.0802001953125, 0.80078125, 0.0322265625, -0.2393798828125, 0.509765625, -0.68408203125, 0.488037109375, -0.26708984375, 0.5693359375, -0.1510009765625, 0.0751953125, 0.1397705078125, -0.94921875, -0.10498046875, 0.022705078125, 0.178466796875, -0.26220703125, -0.404052734375, 0.0472412109375, 1.109375, 0.4423828125, -0.63134765625, -0.233642578125, 0.7197265625, -0.1798095703125, -0.1539306640625, -0.109375, -0.408935546875, 0.23828125, -0.144287109375, -0.47998046875, 0.50732421875, 0.41552734375, -0.370849609375, -0.52392578125, 0.11767578125, 0.1036376953125, -0.115234375, 0.154052734375, -0.262939453125, -0.0689697265625, 0.189208984375, 0.1923828125, 0.35498046875, 0.3798828125, 0.25390625, -0.1466064453125, 0.1435546875, -0.03497314453125, 0.365966796875, 0.3408203125, 0.467529296875, -0.2493896484375, -0.1512451171875, 0.1923828125, -0.258544921875, 0.0537109375, -0.2401123046875, 0.332763671875, 0.36767578125, -0.30419921875, 0.435791015625, -2.130859375, 0.54345703125, 0.484130859375, 0.12255859375, -0.5546875, -0.140380859375, 0.07177734375, -0.155517578125, 0.1298828125, 0.24072265625, 0.375, -0.13525390625, 0.86865234375, -0.33935546875, 0.004638671875, 0.60009765625, 0.2178955078125, 0.27099609375, -0.19482421875, -0.40673828125, 0.2203369140625, 0.49560546875, 0.173095703125, -0.04254150390625, -0.18310546875, 0.29150390625, 0.140625, 0.9892578125, 0.3037109375, 0.357666015625, -0.160400390625, 0.330078125, 0.14111328125, 0.0281982421875, 0.63232421875, -0.71728515625, -0.363525390625, 0.55029296875, -0.5712890625, -0.017333984375, 0.79248046875, -0.43994140625, -0.243896484375, -0.107177734375, -0.44677734375, -0.2783203125, 0.1673583984375, -0.9248046875, -0.06573486328125, 0.0880126953125, -0.0048828125, 0.59619140625, -0.24365234375, 0.349365234375, -0.33642578125, -0.21923828125, -0.206298828125, -0.148193359375, 0.030517578125, -0.194091796875, 0.106201171875, 0.11474609375, 0.099609375, 0.28857421875, -0.208251953125, -0.171875, -0.373779296875, -0.0016689300537109375, 0.49365234375, 1.181640625, -0.1656494140625, -0.274658203125, -0.56103515625, 0.54052734375, 0.40185546875, 0.0948486328125, 0.345703125, -0.24560546875, -1.30078125, -0.1131591796875, 0.023193359375, -0.62353515625, 0.388427734375, 0.5615234375, -0.047637939453125, 0.1708984375, -0.13037109375, 0.09765625, 0.205078125, 0.0457763671875, -0.470703125, 0.263671875, -0.158203125, -0.447021484375, 0.1572265625, -0.134033203125, -0.334716796875, -0.260009765625, -0.2373046875, -0.142333984375, -0.447265625, 0.05517578125, 0.00439453125, -0.48828125, -0.5517578125, -0.19482421875, -0.53857421875, -0.06256103515625, 0.217529296875, -0.06103515625, -0.34716796875, -0.252685546875, -0.156005859375, -0.25439453125, 0.28662109375, 0.49267578125, 0.155517578125, 0.1956787109375, -0.75732421875, -0.346435546875, -0.568359375, -0.227294921875, 0.01416015625, 0.207763671875, 0.1903076171875, 0.110595703125, 0.468994140625, -0.044403076171875, 0.14208984375, 0.015869140625, -0.1656494140625, 0.425537109375, 0.30322265625, 0.61572265625, -0.1453857421875, 0.06640625, -0.7880859375, -0.127685546875, 0.0758056640625, 0.189697265625, -0.240478515625, -0.09185791015625, -0.55078125, 0.27490234375, -1.447265625, 0.14111328125, -0.10076904296875, 1.2470703125, -0.178955078125, 0.37939453125, 0.34521484375, 0.460693359375, -0.12200927734375, -0.04248046875, 0.266357421875, -0.0455322265625, 0.35302734375, 0.63623046875, -0.6552734375, -0.236328125, -0.1429443359375, -0.1495361328125, -0.1165771484375, 0.129638671875, 0.4580078125, 0.02447509765625, -0.6015625, 0.167236328125, -0.19775390625, -0.07537841796875, -0.79296875, 0.841796875, -0.00927734375, -0.0594482421875, 0.01318359375, -0.7744140625, 0.47265625, 0.1551513671875, 0.505859375, -0.006103515625, -0.6435546875, 0.66796875, 0.2822265625, -0.096435546875, 1.1142578125, 0.07952880859375, -0.086181640625, -0.66064453125, -0.2139892578125, -0.240234375, 0.039306640625, 0.135986328125, 0.06024169921875, 0.4072265625, -0.16552734375, 0.11474609375, 0.0140380859375, -0.47705078125, -0.348876953125, 0.35009765625, -0.427734375, -0.2462158203125, 0.09210205078125, -0.409912109375, -0.146484375, -0.060546875, -0.11297607421875, 0.88818359375, 0.0299072265625, -0.27490234375, -0.130126953125, 0.18017578125, 0.1484375, -0.03607177734375, 0.277587890625, -0.1627197265625, 0.251953125, 0.6103515625, -0.47509765625, 0.176025390625, -0.312255859375, -0.84423828125, 0.1575927734375, 0.2376708984375, 0.104248046875, -0.043212890625, 0.05712890625, 0.156982421875, 0.1087646484375, -0.225341796875, 0.416259765625, 0.6796875]}, {"frame_number": 12, "vector": [-0.11572265625, 0.28857421875, -0.0340576171875, -0.1845703125, -0.65234375, 0.33203125, -0.0023193359375, -0.471923828125, -0.1251220703125, -0.00665283203125, -0.13671875, 0.024169921875, 0.31787109375, 0.178466796875, -0.0015869140625, 0.28076171875, 0.140625, -0.144775390625, -0.084716796875, -0.49072265625, -0.55029296875, -6.06640625, -0.323486328125, -0.7490234375, 0.07354736328125, -0.05322265625, -0.12353515625, -0.78173828125, 0.488525390625, -0.201171875, -0.9111328125, -0.59814453125, -0.671875, 0.00067138671875, -0.07403564453125, 0.06549072265625, 0.27978515625, -0.6845703125, -0.0994873046875, -0.42822265625, -0.00262451171875, 0.16552734375, 0.2490234375, 0.090576171875, 0.92529296875, -0.3798828125, -0.38720703125, -0.208251953125, -0.61767578125, -0.19970703125, 0.03460693359375, -0.2509765625, 0.150390625, 0.58203125, 0.188232421875, 0.06573486328125, 0.42822265625, -0.236083984375, 0.2900390625, -0.01312255859375, -0.13232421875, 0.2283935546875, 0.263916015625, 0.181884765625, -0.25732421875, -0.36376953125, -0.513671875, -0.051025390625, 0.261962890625, 0.1009521484375, 0.9453125, -0.0653076171875, -0.023681640625, -0.058837890625, -0.41162109375, 0.82421875, -0.293212890625, -0.46728515625, 0.2314453125, -0.7138671875, 1.23046875, 0.256103515625, 0.5380859375, 0.490478515625, -0.257568359375, 0.0316162109375, 0.145263671875, -0.058013916015625, 0.140625, -0.61181640625, -0.1722412109375, -0.11529541015625, 0.07861328125, -0.1357421875, 0.302490234375, 0.41064453125, 0.0777587890625, 0.55517578125, 0.041534423828125, -0.63525390625, 0.748046875, 0.3525390625, -0.37939453125, -0.638671875, 0.020050048828125, -0.115966796875, -0.019287109375, -0.113525390625, -0.163818359375, -0.025634765625, -0.0504150390625, 0.18994140625, 0.11346435546875, -0.50390625, 0.1351318359375, -0.6171875, -0.2412109375, -0.17626953125, 0.343505859375, 0.533203125, 0.051513671875, -0.2607421875, -0.0401611328125, -0.697265625, -0.1005859375, 0.2392578125, -0.1422119140625, 0.358154296875, 0.162109375, 0.333984375, -0.0404052734375, -0.373779296875, 0.03759765625, 0.02410888671875, 0.599609375, -0.0447998046875, 0.02362060546875, 0.7734375, 0.537109375, 0.002197265625, 0.098876953125, 0.29052734375, 0.1651611328125, -0.0777587890625, -0.208251953125, 0.095458984375, -0.02618408203125, -0.330322265625, 0.1971435546875, 1.0263671875, -0.1739501953125, -1.58203125, 0.0604248046875, -0.36474609375, 0.400146484375, -0.2318115234375, -0.24951171875, 0.955078125, 0.27099609375, -0.41259765625, -0.27490234375, 0.314453125, 0.483154296875, 0.343994140625, -0.318359375, -0.32373046875, 0.01513671875, 0.615234375, 0.0101318359375, 0.0904541015625, 0.1400146484375, 0.449462890625, -0.1905517578125, 0.64453125, -0.273681640625, 1.046875, -1.1552734375, 0.1539306640625, -0.0382080078125, -0.09375, 0.062744140625, 0.17919921875, -0.17236328125, 0.044830322265625, 0.0738525390625, -0.0025634765625, 0.03564453125, 0.2110595703125, -0.0582275390625, 0.14697265625, 0.224609375, 0.169677734375, 0.472900390625, -0.404296875, 0.5517578125, -0.1654052734375, 0.4169921875, -0.20947265625, -0.291015625, 1.2265625, 0.472412109375, -0.0244140625, -0.177001953125, 0.0733642578125, -0.01513671875, 0.049346923828125, 0.63330078125, 0.011962890625, -0.66015625, -0.1533203125, 0.2225341796875, -0.097900390625, 0.239501953125, -0.84423828125, -0.52734375, -0.5, 0.4794921875, 1.02734375, 0.0034942626953125, -0.44384765625, -0.398681640625, -0.54296875, -0.1287841796875, -0.583984375, -0.20263671875, 0.24462890625, -0.1461181640625, -0.24560546875, 0.77783203125, 0.04449462890625, -0.11651611328125, 0.245849609375, 0.099609375, 0.26025390625, -0.361572265625, -0.537109375, 0.755859375, -0.0885009765625, -1.0146484375, -0.0859375, 0.6953125, 0.2391357421875, 0.006591796875, 0.10791015625, -0.2054443359375, 0.9482421875, -0.239501953125, 0.001953125, -0.35693359375, -0.14599609375, -0.077880859375, -0.052490234375, 0.2489013671875, -0.03369140625, -0.65380859375, -0.1077880859375, -0.186279296875, -0.049560546875, 0.463134765625, 0.45263671875, 0.01922607421875, -0.23046875, -0.07470703125, 0.0135498046875, -0.24072265625, 0.01507568359375, -0.76513671875, -0.325927734375, 0.607421875, -0.4248046875, 0.02923583984375, 0.392578125, 0.3271484375, 0.284423828125, -0.2225341796875, -0.720703125, -0.1212158203125, 0.1373291015625, -0.34326171875, -0.372314453125, -0.26123046875, 0.67529296875, -0.4404296875, 0.146728515625, 0.385986328125, 0.318359375, -0.1302490234375, 0.5478515625, -0.556640625, 1.162109375, 0.006561279296875, 0.025054931640625, 0.0672607421875, -0.39697265625, 0.07684326171875, 0.1375732421875, 0.1728515625, 0.383544921875, 0.63916015625, -0.231201171875, -0.015533447265625, 0.7822265625, 0.0245361328125, -0.025146484375, -0.31591796875, 0.388427734375, 0.325439453125, -0.68505859375, 0.10748291015625, -0.0968017578125, 0.392333984375, -0.391357421875, -0.003662109375, -0.333251953125, -0.349365234375, 0.43994140625, -0.09832763671875, 0.000732421875, -0.287109375, 0.19287109375, -0.9130859375, 0.0931396484375, 0.08544921875, -0.442626953125, -0.89990234375, 0.57421875, -0.463134765625, 0.343505859375, -0.55078125, -0.0150146484375, -0.4619140625, -0.0870361328125, -0.1710205078125, -0.459716796875, 0.161865234375, 0.22265625, 0.1307373046875, -0.6201171875, 0.023681640625, 0.0006103515625, -0.73974609375, 0.0543212890625, -0.301025390625, -0.705078125, -0.486083984375, 0.105224609375, -0.435546875, 0.314208984375, 0.210693359375, 0.0306396484375, -0.293701171875, -0.34814453125, 0.263671875, 0.147216796875, 0.42919921875, 0.34375, 0.244384765625, -0.13916015625, -0.337890625, -0.1693115234375, -0.2080078125, -0.0196533203125, 0.48876953125, 0.283203125, -0.60009765625, -0.46826171875, -0.03076171875, 0.269775390625, 0.048828125, -0.08935546875, 0.012603759765625, 0.2666015625, -0.17431640625, 0.311767578125, 0.280517578125, -0.0308837890625, 0.391357421875, -0.22998046875, 0.17236328125, -0.53466796875, -0.02978515625, -0.427490234375, 0.182373046875, 0.66796875, 0.2017822265625, -0.0938720703125, -0.0382080078125, -0.1689453125, -0.279296875, 0.50244140625, -0.6591796875, -0.193603515625, 0.312744140625, 0.26171875, 1.005859375, 0.65771484375, -0.1171875, -0.65771484375, -0.078125, -0.392578125, 0.218505859375, 0.12432861328125, -0.109375, -0.40869140625, -0.1715087890625, 0.02685546875, -0.344970703125, 0.05810546875, 0.3603515625, -0.0030517578125, 0.18505859375, -0.9326171875, -0.71337890625, -0.291015625, -0.36767578125, 0.55029296875, 0.73828125, -0.279296875, 1.12109375, -0.215087890625, 0.2149658203125, -0.053680419921875, -0.215087890625, -0.10888671875, -0.028564453125, -0.48095703125, -1.3876953125, -0.071533203125, 0.09423828125, 0.169677734375, 0.095703125, -6.0625, -0.43310546875, -0.05126953125, -0.7890625, -0.166259765625, -0.1029052734375, -1.2109375, -0.6376953125, -0.207275390625, 0.0087890625, 0.57275390625, 0.66015625, -0.2236328125, 0.03314208984375, 0.1256103515625, 0.220458984375, 0.005615234375, -0.0411376953125, 0.202880859375, 0.049560546875, 0.21484375, -0.1107177734375, 0.077880859375, 0.436279296875, -0.28515625, -0.2413330078125, 0.372314453125, -0.66015625, -0.615234375, -0.093505859375, 0.04443359375, -0.30029296875, -0.3349609375, -0.40478515625, -0.0552978515625, 0.57666015625, 0.79443359375, -0.3193359375, -0.400390625, 0.197998046875, -0.98876953125, 0.360595703125, -0.04638671875, 0.0147705078125, 0.240478515625, 0.319091796875, -0.11370849609375, -0.7666015625, -0.73681640625, 0.132080078125, 0.1668701171875, -1.0234375, -0.0205078125, 0.3427734375, -0.045654296875, 0.246337890625, -0.032867431640625, 0.10589599609375, 0.182861328125, 0.185791015625, 0.56689453125, 0.338623046875, 0.4345703125, -0.380859375, -0.0107421875, 0.046173095703125, 0.011962890625, 0.10888671875, 0.075927734375, -0.1201171875, 0.02984619140625, 0.68115234375, -0.158447265625, -0.26806640625, -0.123291015625, -0.10546875, 0.062255859375, -0.19677734375, 0.070068359375, 0.2098388671875, 0.058349609375, 0.06866455078125, 0.33203125, -0.246337890625, -0.1846923828125, -0.094970703125, -0.3544921875, -0.5830078125, -0.0828857421875, 0.693359375, 0.16357421875, -0.232177734375, 0.10205078125, 0.201171875, 1.2900390625, 0.1826171875, 0.1363525390625, 0.1507568359375, 0.0782470703125, -0.3525390625, -0.44580078125, -0.06982421875, -0.087646484375, 0.338134765625, 0.272216796875, 0.0556640625, 0.013916015625, 0.42041015625, -0.0638427734375, 0.7412109375, 0.1524658203125, 0.39990234375, 0.55810546875, -0.1768798828125, 0.51318359375, -1.53125, -0.409423828125, -0.25048828125, 0.102294921875, -0.0869140625, -0.315673828125, 0.77490234375, 0.296142578125, -0.1146240234375, 0.079833984375, 0.274169921875, -0.0948486328125, -0.222900390625, -0.9345703125, -0.42138671875, -0.2861328125, -0.4560546875, -0.9365234375, 0.28662109375, 0.28857421875, 0.473876953125, -0.40771484375, -0.0645751953125, 0.0274658203125, -0.070556640625, 0.3681640625, -0.25390625, 0.11474609375, 0.1463623046875, 0.439453125, -0.369873046875, 0.18994140625, -0.345703125, -0.478271484375, -0.3603515625, 0.1888427734375, -0.148681640625, -0.1011962890625, 0.05548095703125, -0.222412109375, 0.0224609375, 0.45654296875, -0.325439453125, 0.20263671875, 0.352294921875, 0.00138092041015625, -0.236572265625, 0.047607421875, 0.0927734375, 0.04571533203125, -0.087646484375, -0.2215576171875, -1.521484375, -0.44970703125, 0.140625, -0.0587158203125, -0.333251953125, -0.3798828125, -0.24951171875, -0.137451171875, -0.09375, -0.265625, -0.214599609375, -0.84912109375, 0.224365234375, 0.35009765625, -0.2269287109375, -0.221923828125, -0.1329345703125, -0.591796875, 0.21240234375, 0.2919921875, -0.051727294921875, -0.2435302734375, 0.02215576171875, -0.06640625, 0.04803466796875, 0.04632568359375, -0.25146484375, -0.01428985595703125, 0.066162109375, 0.43603515625, 0.244384765625, -0.075439453125, 0.124267578125, 0.02728271484375, -0.385009765625, 0.71826171875, 0.051025390625, -0.119140625, -0.274169921875, 1.328125, 0.034423828125, -0.30126953125, 0.01416015625, -0.424072265625, 0.348388671875, -0.242919921875, -0.1923828125, 0.32177734375, 0.1669921875, -0.04052734375, 0.270263671875, 0.1529541015625, 0.262939453125, -0.330078125, 0.39111328125, 0.23193359375, -0.6083984375, 0.1826171875, 0.2978515625, -0.112060546875, -0.23681640625, -0.2276611328125, -0.276123046875, -0.83837890625, -0.008056640625, -0.14501953125, -0.07574462890625, -0.305419921875, 0.0673828125, -0.104248046875, 0.4677734375, -0.45947265625, -1.3828125, -0.0672607421875, -0.149169921875, 0.224853515625, 0.21923828125, -0.114013671875, 0.3154296875, 0.1470947265625, -0.15478515625, 0.28173828125, 0.0147705078125, 0.1336669921875, 0.44482421875, -0.00341796875, 0.1488037109375, -0.1650390625, 0.326171875, 0.26513671875, -0.3828125, -0.6240234375, -0.17431640625, -0.47265625, -0.0048828125, 0.296875, -0.26513671875, 0.8466796875, -0.0550537109375, 0.11602783203125, -0.84130859375, -0.03509521484375, -0.017578125, 0.193359375, -0.583984375, 0.320556640625, -0.54833984375, 0.4794921875, -0.009979248046875, 0.198974609375, -0.119140625, -0.52490234375, -0.751953125, 0.10986328125, -0.44921875, 0.457275390625, -0.30126953125, 0.9892578125, -0.376220703125, 0.0565185546875, -0.11474609375, -0.0428466796875, -0.42333984375, 0.092529296875, 1.4443359375, 0.044677734375, 0.2498779296875, -0.0657958984375, 0.407958984375, -1.0927734375, -0.26513671875, 0.509765625, 0.06298828125, -0.2734375, 0.498291015625, 0.18701171875, 0.097412109375, 0.1689453125, 0.0577392578125, 0.197021484375, -0.058349609375, 0.43603515625, -0.443115234375, 0.010009765625, -0.6787109375, 0.0421142578125, 0.100830078125, -0.46044921875, -0.00341796875, -0.11651611328125, -0.146484375, 0.0618896484375, -0.89453125, -0.2666015625, 0.23583984375, -0.0665283203125, 0.2230224609375, -0.054229736328125, -0.109130859375, -0.78369140625, 0.0362548828125, 0.294189453125, -0.07379150390625, 0.06353759765625, -0.233642578125, -0.54638671875, 0.572265625, -0.439697265625, -0.01263427734375, -0.615234375, 0.01593017578125, -0.367431640625, 0.09637451171875, -0.04345703125, -0.0233154296875, 0.11566162109375, 0.2314453125, -2.22265625, 0.35009765625, -0.26123046875, 0.2763671875, 0.65478515625, -0.067626953125, -0.315185546875, -0.23681640625, 0.1728515625, -0.099609375, 0.251953125, 0.021240234375, 0.08251953125, 0.46875, 0.4873046875, 0.38818359375, -0.000244140625, 0.267578125, 0.705078125, -0.291259765625, 0.367431640625, 0.0592041015625, 0.1317138671875, -0.227294921875, 0.101806640625, 0.2919921875, 0.26708984375, -0.76171875, 0.59765625, 0.1522216796875, -0.1224365234375, 0.012939453125, -0.49853515625, 0.0064697265625, -0.007080078125, 0.2958984375, -0.29248046875, -0.07135009765625, -0.19775390625, -0.216796875, 0.47705078125, -0.040771484375, -0.31396484375, -0.66796875, -0.38916015625, 0.282470703125, 0.100830078125, -0.196044921875, 0.4111328125, -0.11553955078125, 0.306396484375, -0.4140625, -0.247314453125, 0.167724609375, -0.34228515625, -0.50390625, -0.6494140625, -0.6171875, -0.0374755859375, 0.7607421875, 0.63818359375, -0.230712890625, 0.0654296875, 0.7060546875, -0.650390625, 0.288818359375, -0.67822265625, -0.60009765625, -0.517578125, 0.0845947265625, 0.08551025390625, -0.40380859375, -0.313232421875, 0.1651611328125, -0.13818359375, 0.14208984375, 0.44140625, 0.0231170654296875, 0.1524658203125, 0.024688720703125, -0.439453125, 0.09130859375, -0.0426025390625, -0.0953369140625, 0.83447265625, 0.02490234375, -0.261962890625, 0.4970703125, -0.65771484375, 0.48193359375, -0.285400390625, 0.578125, -0.1632080078125, 0.106689453125, 0.1180419921875, -0.96923828125, -0.140625, 0.02392578125, 0.1475830078125, -0.29248046875, -0.405029296875, 0.05364990234375, 1.111328125, 0.451904296875, -0.62158203125, -0.2734375, 0.724609375, -0.1849365234375, -0.174072265625, -0.11798095703125, -0.4072265625, 0.25537109375, -0.1712646484375, -0.4365234375, 0.46240234375, 0.40185546875, -0.328125, -0.52783203125, 0.1260986328125, 0.084228515625, -0.0845947265625, 0.1259765625, -0.27392578125, -0.07763671875, 0.183837890625, 0.2330322265625, 0.37548828125, 0.38916015625, 0.265380859375, -0.133056640625, 0.16015625, -0.04986572265625, 0.3837890625, 0.347412109375, 0.53271484375, -0.27490234375, -0.11126708984375, 0.1895751953125, -0.283203125, 0.02783203125, -0.2247314453125, 0.3603515625, 0.394775390625, -0.29931640625, 0.468994140625, -2.134765625, 0.5419921875, 0.45458984375, 0.127685546875, -0.53759765625, -0.1697998046875, 0.0921630859375, -0.131591796875, 0.123291015625, 0.22119140625, 0.38671875, -0.155517578125, 0.82666015625, -0.338134765625, -0.0025634765625, 0.62548828125, 0.245849609375, 0.2734375, -0.2158203125, -0.38623046875, 0.217529296875, 0.48974609375, 0.1593017578125, -0.030853271484375, -0.1939697265625, 0.266357421875, 0.1533203125, 1.01171875, 0.312744140625, 0.348388671875, -0.18212890625, 0.344482421875, 0.1767578125, 0.01220703125, 0.63037109375, -0.732421875, -0.315673828125, 0.53369140625, -0.603515625, 0.05029296875, 0.8349609375, -0.45751953125, -0.269287109375, -0.1103515625, -0.42431640625, -0.232177734375, 0.111572265625, -0.935546875, -0.05810546875, 0.10882568359375, -0.0289306640625, 0.6494140625, -0.27392578125, 0.3720703125, -0.343994140625, -0.22998046875, -0.1953125, -0.112548828125, 0.0390625, -0.158447265625, 0.09375, 0.1009521484375, 0.081298828125, 0.252685546875, -0.205322265625, -0.171630859375, -0.40771484375, 0.0167083740234375, 0.48095703125, 1.25, -0.1549072265625, -0.2548828125, -0.62353515625, 0.57080078125, 0.39306640625, 0.053955078125, 0.3486328125, -0.259765625, -1.318359375, -0.09027099609375, 0.02734375, -0.60498046875, 0.411376953125, 0.51171875, -0.05474853515625, 0.16796875, -0.158447265625, 0.101318359375, 0.17138671875, 0.052734375, -0.47314453125, 0.236083984375, -0.143310546875, -0.4619140625, 0.1414794921875, -0.15087890625, -0.319091796875, -0.256103515625, -0.191162109375, -0.171630859375, -0.47607421875, 0.01513671875, 0.02099609375, -0.443359375, -0.53955078125, -0.158447265625, -0.47998046875, -0.0462646484375, 0.19384765625, -0.077880859375, -0.31494140625, -0.22265625, -0.15673828125, -0.2376708984375, 0.24755859375, 0.529296875, 0.1485595703125, 0.2432861328125, -0.74951171875, -0.346435546875, -0.50048828125, -0.2333984375, 0.020263671875, 0.2017822265625, 0.1890869140625, 0.10302734375, 0.470947265625, -0.03753662109375, 0.1390380859375, 0.0169677734375, -0.154296875, 0.437744140625, 0.24365234375, 0.580078125, -0.134033203125, 0.0634765625, -0.755859375, -0.0914306640625, 0.0703125, 0.21435546875, -0.26708984375, -0.10833740234375, -0.5712890625, 0.262939453125, -1.439453125, 0.14453125, -0.09320068359375, 1.2412109375, -0.15966796875, 0.37939453125, 0.3525390625, 0.458984375, -0.119140625, -0.008056640625, 0.284423828125, -0.08349609375, 0.3720703125, 0.59326171875, -0.6640625, -0.189453125, -0.14453125, -0.15673828125, -0.09027099609375, 0.1787109375, 0.4736328125, 0.01806640625, -0.58544921875, 0.182861328125, -0.21240234375, -0.05609130859375, -0.78173828125, 0.81640625, -0.02783203125, -0.072021484375, -0.0069580078125, -0.7529296875, 0.463134765625, 0.2191162109375, 0.469482421875, -0.019775390625, -0.6552734375, 0.66943359375, 0.25439453125, -0.08026123046875, 1.0859375, 0.0335693359375, -0.04931640625, -0.7099609375, -0.22509765625, -0.238037109375, 0.0035400390625, 0.08349609375, 0.0911865234375, 0.3896484375, -0.184326171875, 0.1298828125, 0.0101318359375, -0.5126953125, -0.323486328125, 0.30517578125, -0.43798828125, -0.26416015625, 0.109375, -0.38671875, -0.11407470703125, -0.08929443359375, -0.099609375, 0.849609375, -0.036376953125, -0.2421875, -0.124755859375, 0.1707763671875, 0.1185302734375, -0.03179931640625, 0.29150390625, -0.150390625, 0.21435546875, 0.62548828125, -0.4111328125, 0.1357421875, -0.2939453125, -0.84033203125, 0.1697998046875, 0.279296875, 0.092529296875, -0.056793212890625, 0.065673828125, 0.12353515625, 0.0960693359375, -0.223388671875, 0.4541015625, 0.7255859375]}, {"frame_number": 13, "vector": [-0.153076171875, 0.266845703125, -0.020263671875, -0.1832275390625, -0.6357421875, 0.30029296875, -0.036285400390625, -0.46875, -0.12841796875, -0.0016937255859375, -0.149658203125, 0.05859375, 0.32568359375, 0.16845703125, -0.03369140625, 0.274658203125, 0.132080078125, -0.172607421875, -0.10418701171875, -0.46142578125, -0.5390625, -6.0546875, -0.31689453125, -0.72998046875, 0.07684326171875, -0.03411865234375, -0.177734375, -0.796875, 0.51953125, -0.19677734375, -0.89453125, -0.61669921875, -0.67041015625, -0.00372314453125, -0.1112060546875, 0.022705078125, 0.3037109375, -0.6748046875, -0.06488037109375, -0.399658203125, 0.001953125, 0.19140625, 0.263671875, 0.064453125, 0.90087890625, -0.37255859375, -0.404052734375, -0.2325439453125, -0.599609375, -0.2001953125, 0.0305938720703125, -0.25927734375, 0.1689453125, 0.59130859375, 0.189208984375, 0.04595947265625, 0.4130859375, -0.241943359375, 0.255615234375, -0.01129150390625, -0.14208984375, 0.2154541015625, 0.282958984375, 0.209228515625, -0.2587890625, -0.35009765625, -0.50927734375, -0.0263671875, 0.239990234375, 0.046142578125, 0.95068359375, -0.0550537109375, 0.00439453125, -0.090576171875, -0.443603515625, 0.82861328125, -0.242431640625, -0.4599609375, 0.2490234375, -0.736328125, 1.23046875, 0.248291015625, 0.5478515625, 0.5048828125, -0.30517578125, 0.0218505859375, 0.139404296875, -0.061126708984375, 0.16943359375, -0.5537109375, -0.182373046875, -0.1297607421875, 0.064697265625, -0.12060546875, 0.294189453125, 0.414306640625, 0.05810546875, 0.55712890625, 0.01047515869140625, -0.66064453125, 0.728515625, 0.3525390625, -0.380859375, -0.6552734375, 0.019805908203125, -0.1090087890625, 0.00592041015625, -0.0792236328125, -0.16259765625, 0.020263671875, -0.06072998046875, 0.197021484375, 0.11358642578125, -0.47607421875, 0.1610107421875, -0.650390625, -0.238037109375, -0.1630859375, 0.359375, 0.53662109375, 0.057861328125, -0.258544921875, -0.004150390625, -0.6943359375, -0.0811767578125, 0.205078125, -0.103271484375, 0.3701171875, 0.1859130859375, 0.311767578125, -0.02685546875, -0.36669921875, 0.07196044921875, 0.0460205078125, 0.5947265625, -0.044677734375, 0.020751953125, 0.75634765625, 0.5263671875, 0.016845703125, 0.07373046875, 0.251708984375, 0.1583251953125, -0.0546875, -0.2373046875, 0.096923828125, -0.03778076171875, -0.3388671875, 0.19970703125, 1.06640625, -0.155029296875, -1.5625, 0.048095703125, -0.3505859375, 0.391357421875, -0.230224609375, -0.2578125, 0.9248046875, 0.26318359375, -0.40478515625, -0.273193359375, 0.3115234375, 0.464111328125, 0.3564453125, -0.29931640625, -0.29052734375, -0.0037841796875, 0.64111328125, -0.0123291015625, 0.0858154296875, 0.1719970703125, 0.380859375, -0.1978759765625, 0.66796875, -0.23193359375, 1.013671875, -1.1953125, 0.15185546875, -0.064453125, -0.092041015625, 0.048095703125, 0.197021484375, -0.17919921875, 0.026214599609375, 0.0804443359375, 0.0274658203125, 0.071533203125, 0.23876953125, -0.044677734375, 0.142333984375, 0.225830078125, 0.1275634765625, 0.461669921875, -0.36279296875, 0.5693359375, -0.193359375, 0.431396484375, -0.2197265625, -0.29931640625, 1.173828125, 0.46875, -0.0087890625, -0.22216796875, 0.072021484375, 0.00927734375, 0.0521240234375, 0.66748046875, 0.044464111328125, -0.71630859375, -0.12164306640625, 0.2059326171875, -0.08544921875, 0.2392578125, -0.8525390625, -0.479248046875, -0.51611328125, 0.53173828125, 0.9921875, 0.024017333984375, -0.44677734375, -0.40771484375, -0.54248046875, -0.154052734375, -0.59228515625, -0.220703125, 0.25634765625, -0.182373046875, -0.26611328125, 0.82861328125, 0.0252685546875, -0.130615234375, 0.251953125, 0.06787109375, 0.245849609375, -0.275390625, -0.5419921875, 0.7568359375, -0.0819091796875, -1.03515625, -0.072509765625, 0.7138671875, 0.2376708984375, -0.03448486328125, 0.109375, -0.193603515625, 0.935546875, -0.28564453125, -0.01904296875, -0.3701171875, -0.176025390625, -0.0576171875, -0.0352783203125, 0.2568359375, -0.053955078125, -0.66845703125, -0.073486328125, -0.183349609375, -0.06396484375, 0.43359375, 0.474609375, 0.00823974609375, -0.210693359375, -0.08740234375, 0.02294921875, -0.233154296875, 0.0338134765625, -0.78564453125, -0.31884765625, 0.62548828125, -0.453369140625, 0.0816650390625, 0.377685546875, 0.312255859375, 0.3037109375, -0.2191162109375, -0.69580078125, -0.1378173828125, 0.150146484375, -0.319091796875, -0.392333984375, -0.2392578125, 0.6923828125, -0.449462890625, 0.13916015625, 0.41748046875, 0.326171875, -0.145751953125, 0.5478515625, -0.53759765625, 1.15625, -0.0191497802734375, 0.0264892578125, 0.093017578125, -0.373779296875, 0.06903076171875, 0.13671875, 0.182861328125, 0.423828125, 0.58056640625, -0.248779296875, -0.002227783203125, 0.76904296875, 0.044830322265625, -0.038330078125, -0.345947265625, 0.373291015625, 0.27978515625, -0.67431640625, 0.1182861328125, -0.138427734375, 0.3671875, -0.431640625, -0.0048828125, -0.3388671875, -0.3408203125, 0.45947265625, -0.1002197265625, 0.016357421875, -0.2734375, 0.2333984375, -0.8828125, 0.0810546875, 0.04351806640625, -0.43212890625, -0.86669921875, 0.5869140625, -0.462890625, 0.358642578125, -0.5478515625, -0.00982666015625, -0.466796875, -0.119384765625, -0.193603515625, -0.4072265625, 0.1585693359375, 0.1865234375, 0.1409912109375, -0.619140625, -0.0174560546875, -0.015625, -0.701171875, 0.0340576171875, -0.3193359375, -0.77197265625, -0.48779296875, 0.126953125, -0.443115234375, 0.265380859375, 0.20166015625, 0.052490234375, -0.326416015625, -0.38818359375, 0.2138671875, 0.13232421875, 0.3984375, 0.344970703125, 0.2490234375, -0.1478271484375, -0.32275390625, -0.1466064453125, -0.1871337890625, 0.0181884765625, 0.50341796875, 0.300537109375, -0.63232421875, -0.464599609375, -0.054931640625, 0.27001953125, 0.0955810546875, -0.08837890625, -0.0357666015625, 0.275390625, -0.19140625, 0.306640625, 0.31298828125, -0.04736328125, 0.40087890625, -0.2349853515625, 0.19580078125, -0.53271484375, -0.048828125, -0.40869140625, 0.152587890625, 0.67626953125, 0.2176513671875, -0.0947265625, -0.05194091796875, -0.14208984375, -0.323486328125, 0.5498046875, -0.65576171875, -0.20947265625, 0.290283203125, 0.232666015625, 1.01953125, 0.69580078125, -0.154541015625, -0.662109375, -0.0543212890625, -0.406982421875, 0.255859375, 0.12335205078125, -0.114501953125, -0.43115234375, -0.139892578125, 0.020263671875, -0.33740234375, 0.0972900390625, 0.35791015625, -0.025146484375, 0.25634765625, -0.8896484375, -0.7216796875, -0.32275390625, -0.4013671875, 0.54833984375, 0.734375, -0.29833984375, 1.150390625, -0.202392578125, 0.18408203125, -0.061767578125, -0.189453125, -0.143310546875, -0.12152099609375, -0.46435546875, -1.384765625, -0.06646728515625, 0.12158203125, 0.210205078125, 0.133056640625, -6.06640625, -0.451171875, -0.0732421875, -0.744140625, -0.194580078125, -0.0875244140625, -1.19921875, -0.65869140625, -0.218017578125, 0.0194091796875, 0.5458984375, 0.673828125, -0.21875, 0.03399658203125, 0.2154541015625, 0.26318359375, 0.0140380859375, -0.00732421875, 0.2071533203125, 0.0618896484375, 0.209228515625, -0.13720703125, 0.096923828125, 0.471435546875, -0.286865234375, -0.274658203125, 0.36572265625, -0.66650390625, -0.65576171875, -0.06103515625, 0.09375, -0.30224609375, -0.35302734375, -0.4189453125, -0.0699462890625, 0.56396484375, 0.80078125, -0.330078125, -0.37109375, 0.186279296875, -0.99365234375, 0.32275390625, -0.0418701171875, 0.02197265625, 0.240234375, 0.345947265625, -0.08160400390625, -0.7275390625, -0.75, 0.153076171875, 0.1676025390625, -1.0380859375, -0.02490234375, 0.368896484375, -0.02825927734375, 0.2392578125, -0.065673828125, 0.14013671875, 0.1820068359375, 0.19091796875, 0.58740234375, 0.353515625, 0.44970703125, -0.3544921875, 0.004150390625, 0.061676025390625, 0.0126953125, 0.09912109375, 0.0770263671875, -0.0836181640625, 0.03521728515625, 0.6572265625, -0.1614990234375, -0.283203125, -0.1064453125, -0.11859130859375, 0.0643310546875, -0.213623046875, 0.0634765625, 0.195556640625, 0.0587158203125, 0.07415771484375, 0.29736328125, -0.224609375, -0.1793212890625, -0.0830078125, -0.36767578125, -0.59130859375, -0.121337890625, 0.69921875, 0.126220703125, -0.19921875, 0.1002197265625, 0.1846923828125, 1.2724609375, 0.1943359375, 0.101806640625, 0.15283203125, 0.04541015625, -0.38134765625, -0.475341796875, -0.06787109375, -0.1005859375, 0.306884765625, 0.2666015625, 0.07568359375, -0.00018310546875, 0.42041015625, -0.0711669921875, 0.7890625, 0.1495361328125, 0.417724609375, 0.5478515625, -0.1568603515625, 0.5185546875, -1.517578125, -0.4365234375, -0.212890625, 0.1014404296875, -0.072021484375, -0.356689453125, 0.8251953125, 0.310302734375, -0.11376953125, 0.0869140625, 0.289306640625, -0.0943603515625, -0.220947265625, -0.9326171875, -0.414306640625, -0.257080078125, -0.459228515625, -0.96630859375, 0.271484375, 0.275634765625, 0.50244140625, -0.42919921875, -0.04644775390625, 0.049560546875, -0.087158203125, 0.37548828125, -0.2578125, 0.0867919921875, 0.157958984375, 0.408935546875, -0.364013671875, 0.1861572265625, -0.3779296875, -0.533203125, -0.320068359375, 0.2052001953125, -0.135009765625, -0.0714111328125, 0.040679931640625, -0.176025390625, -0.015625, 0.44384765625, -0.343505859375, 0.177001953125, 0.35595703125, 0.01422882080078125, -0.26318359375, 0.0662841796875, 0.09844970703125, 0.028076171875, -0.0706787109375, -0.219970703125, -1.4931640625, -0.4541015625, 0.1407470703125, -0.0452880859375, -0.33935546875, -0.3564453125, -0.2861328125, -0.1300048828125, -0.08990478515625, -0.284423828125, -0.24853515625, -0.84375, 0.1773681640625, 0.3515625, -0.21240234375, -0.205078125, -0.1451416015625, -0.5966796875, 0.216796875, 0.259765625, -0.0623779296875, -0.218994140625, -0.02178955078125, -0.03955078125, 0.037567138671875, 0.0540771484375, -0.24169921875, 0.0049591064453125, 0.0657958984375, 0.486083984375, 0.2890625, -0.0682373046875, 0.15234375, -0.0169830322265625, -0.4072265625, 0.72119140625, 0.04534912109375, -0.122314453125, -0.283935546875, 1.326171875, 0.01611328125, -0.326904296875, -0.0001220703125, -0.39892578125, 0.34521484375, -0.260986328125, -0.20263671875, 0.346435546875, 0.1527099609375, -0.037200927734375, 0.232177734375, 0.140625, 0.2469482421875, -0.33642578125, 0.39306640625, 0.218505859375, -0.62255859375, 0.213623046875, 0.32177734375, -0.08599853515625, -0.23046875, -0.251953125, -0.294921875, -0.91162109375, -0.0174560546875, -0.1220703125, -0.04876708984375, -0.30517578125, 0.072265625, -0.140869140625, 0.4423828125, -0.45068359375, -1.423828125, -0.102294921875, -0.13330078125, 0.239501953125, 0.20751953125, -0.1275634765625, 0.244140625, 0.156005859375, -0.117431640625, 0.31884765625, -0.0020751953125, 0.105712890625, 0.455810546875, 0.03125, 0.1207275390625, -0.1435546875, 0.33544921875, 0.231201171875, -0.379150390625, -0.62646484375, -0.19384765625, -0.4697265625, 0.02734375, 0.2890625, -0.232666015625, 0.8427734375, -0.0517578125, 0.083251953125, -0.84228515625, -0.00286865234375, -0.0107421875, 0.223876953125, -0.5947265625, 0.322265625, -0.5322265625, 0.482421875, -0.01849365234375, 0.18603515625, -0.1173095703125, -0.5390625, -0.7548828125, 0.107421875, -0.464111328125, 0.43505859375, -0.326416015625, 1.005859375, -0.39013671875, 0.0648193359375, -0.11279296875, -0.0504150390625, -0.43896484375, 0.1240234375, 1.421875, 0.03204345703125, 0.2257080078125, -0.07080078125, 0.3818359375, -1.025390625, -0.2626953125, 0.529296875, 0.07269287109375, -0.2376708984375, 0.498779296875, 0.196044921875, 0.083740234375, 0.155029296875, 0.0511474609375, 0.21923828125, -0.04443359375, 0.40625, -0.433837890625, -0.00146484375, -0.6630859375, 0.044677734375, 0.1156005859375, -0.453125, -0.0032958984375, -0.148193359375, -0.1630859375, 0.12255859375, -0.8642578125, -0.259765625, 0.2666015625, -0.0911865234375, 0.17333984375, -0.054412841796875, -0.0987548828125, -0.7822265625, 0.0098876953125, 0.29931640625, -0.084228515625, 0.0931396484375, -0.235595703125, -0.5283203125, 0.58154296875, -0.4453125, 0.010009765625, -0.61083984375, 0.03076171875, -0.33984375, 0.093017578125, -0.00927734375, -0.02734375, 0.14990234375, 0.244140625, -2.279296875, 0.30712890625, -0.3115234375, 0.26513671875, 0.65673828125, -0.098876953125, -0.31103515625, -0.263671875, 0.1708984375, -0.1171875, 0.26123046875, 0.053466796875, 0.0760498046875, 0.451416015625, 0.49365234375, 0.3916015625, -0.0079345703125, 0.252685546875, 0.71142578125, -0.30126953125, 0.3837890625, 0.08551025390625, 0.1590576171875, -0.237548828125, 0.116455078125, 0.283935546875, 0.3046875, -0.763671875, 0.56982421875, 0.101806640625, -0.0870361328125, -0.0042724609375, -0.55224609375, -0.01513671875, 0.0008544921875, 0.302734375, -0.300048828125, -0.08050537109375, -0.199951171875, -0.1787109375, 0.5029296875, -0.0675048828125, -0.28369140625, -0.6259765625, -0.40234375, 0.27294921875, 0.0865478515625, -0.166748046875, 0.385986328125, -0.09649658203125, 0.314208984375, -0.414794921875, -0.25146484375, 0.1590576171875, -0.293701171875, -0.509765625, -0.64794921875, -0.5703125, -0.06024169921875, 0.765625, 0.673828125, -0.21630859375, 0.04803466796875, 0.681640625, -0.67138671875, 0.2978515625, -0.716796875, -0.583984375, -0.49072265625, 0.0709228515625, 0.059967041015625, -0.390869140625, -0.28759765625, 0.18408203125, -0.16357421875, 0.15478515625, 0.459716796875, 0.035736083984375, 0.16552734375, 0.021728515625, -0.453125, 0.10302734375, -0.074951171875, -0.115478515625, 0.83203125, 0.059326171875, -0.2498779296875, 0.485107421875, -0.66162109375, 0.48095703125, -0.289794921875, 0.57177734375, -0.1593017578125, 0.062255859375, 0.1358642578125, -0.99365234375, -0.1337890625, 0.0267333984375, 0.1551513671875, -0.29541015625, -0.425048828125, 0.05450439453125, 1.115234375, 0.45751953125, -0.6083984375, -0.245849609375, 0.74853515625, -0.167724609375, -0.130615234375, -0.0762939453125, -0.38916015625, 0.253173828125, -0.179931640625, -0.476806640625, 0.471923828125, 0.40625, -0.358642578125, -0.59716796875, 0.1358642578125, 0.1138916015625, -0.09765625, 0.1219482421875, -0.285888671875, -0.06640625, 0.17138671875, 0.238525390625, 0.361083984375, 0.41943359375, 0.214599609375, -0.154541015625, 0.14501953125, -0.0885009765625, 0.3935546875, 0.3671875, 0.5283203125, -0.2646484375, -0.1727294921875, 0.154052734375, -0.251953125, -0.00927734375, -0.230224609375, 0.376220703125, 0.386474609375, -0.27490234375, 0.446533203125, -2.1328125, 0.57666015625, 0.4794921875, 0.12109375, -0.5517578125, -0.15869140625, 0.090576171875, -0.15966796875, 0.10955810546875, 0.23046875, 0.386962890625, -0.150634765625, 0.85009765625, -0.312255859375, -0.0184326171875, 0.61669921875, 0.255859375, 0.27587890625, -0.1806640625, -0.39306640625, 0.2213134765625, 0.45849609375, 0.1600341796875, -0.024932861328125, -0.2109375, 0.24072265625, 0.169677734375, 0.9765625, 0.2646484375, 0.334716796875, -0.1650390625, 0.349609375, 0.152587890625, 0.01898193359375, 0.65283203125, -0.779296875, -0.317138671875, 0.56005859375, -0.58056640625, 0.04541015625, 0.8330078125, -0.439453125, -0.26025390625, -0.0733642578125, -0.4033203125, -0.248046875, 0.1107177734375, -0.966796875, -0.03924560546875, 0.11376953125, -0.015869140625, 0.63916015625, -0.27392578125, 0.3623046875, -0.373291015625, -0.2548828125, -0.1885986328125, -0.11962890625, 0.037109375, -0.1650390625, 0.1031494140625, 0.074951171875, 0.0782470703125, 0.25048828125, -0.194580078125, -0.146484375, -0.375732421875, 0.016448974609375, 0.52001953125, 1.275390625, -0.188232421875, -0.259765625, -0.599609375, 0.54736328125, 0.38916015625, 0.068603515625, 0.3447265625, -0.2900390625, -1.3017578125, -0.08489990234375, 0.013427734375, -0.619140625, 0.38671875, 0.5478515625, -0.085205078125, 0.16015625, -0.1634521484375, 0.1142578125, 0.18896484375, 0.0294189453125, -0.4541015625, 0.23291015625, -0.1309814453125, -0.42578125, 0.1513671875, -0.1387939453125, -0.31982421875, -0.248779296875, -0.18896484375, -0.147705078125, -0.461181640625, 0.045166015625, 0.007080078125, -0.43115234375, -0.5166015625, -0.17578125, -0.4619140625, -0.0155029296875, 0.1787109375, -0.058837890625, -0.301513671875, -0.2353515625, -0.158203125, -0.248779296875, 0.25048828125, 0.490234375, 0.1329345703125, 0.2393798828125, -0.732421875, -0.3603515625, -0.53515625, -0.1988525390625, 0.0084228515625, 0.206298828125, 0.178466796875, 0.0787353515625, 0.45654296875, -0.038116455078125, 0.17333984375, -0.03466796875, -0.1956787109375, 0.39990234375, 0.281005859375, 0.5986328125, -0.190673828125, 0.07470703125, -0.78662109375, -0.1114501953125, 0.0599365234375, 0.2171630859375, -0.280029296875, -0.10174560546875, -0.54296875, 0.279052734375, -1.419921875, 0.13037109375, -0.10467529296875, 1.24609375, -0.154541015625, 0.38525390625, 0.358642578125, 0.44482421875, -0.1173095703125, -0.04486083984375, 0.295654296875, -0.091552734375, 0.38134765625, 0.6064453125, -0.658203125, -0.192626953125, -0.158203125, -0.1697998046875, -0.11767578125, 0.146484375, 0.48193359375, 0.018310546875, -0.6044921875, 0.1728515625, -0.219970703125, -0.0292816162109375, -0.7734375, 0.861328125, 0.02734375, -0.047119140625, -0.0206298828125, -0.7646484375, 0.449462890625, 0.1846923828125, 0.52197265625, -0.015625, -0.66943359375, 0.68212890625, 0.255859375, -0.10321044921875, 1.1064453125, 0.01806640625, -0.040771484375, -0.7119140625, -0.236083984375, -0.25830078125, -0.0076904296875, 0.103515625, 0.0621337890625, 0.45458984375, -0.168212890625, 0.132080078125, -0.0076904296875, -0.51513671875, -0.351318359375, 0.29296875, -0.43310546875, -0.269287109375, 0.1239013671875, -0.38330078125, -0.142578125, -0.07708740234375, -0.1055908203125, 0.8525390625, -0.031494140625, -0.25439453125, -0.136962890625, 0.193115234375, 0.1634521484375, -0.053955078125, 0.331787109375, -0.1580810546875, 0.2265625, 0.669921875, -0.4384765625, 0.186279296875, -0.292236328125, -0.84228515625, 0.1978759765625, 0.2724609375, 0.11572265625, -0.036376953125, 0.04345703125, 0.14404296875, 0.128173828125, -0.20556640625, 0.42822265625, 0.72412109375]}, {"frame_number": 14, "vector": [-0.144287109375, 0.26806640625, -0.00927734375, -0.1824951171875, -0.64892578125, 0.305908203125, 0.00177001953125, -0.46728515625, -0.128173828125, 0.0032196044921875, -0.1572265625, 0.0537109375, 0.34326171875, 0.1591796875, -0.0028076171875, 0.283447265625, 0.13623046875, -0.216796875, -0.142333984375, -0.474609375, -0.5537109375, -6.078125, -0.329345703125, -0.744140625, 0.05426025390625, -0.0269775390625, -0.16943359375, -0.82568359375, 0.501953125, -0.220947265625, -0.9306640625, -0.59814453125, -0.66357421875, -0.011383056640625, -0.09576416015625, 0.023681640625, 0.28662109375, -0.6533203125, -0.05633544921875, -0.403564453125, 0.0081787109375, 0.178466796875, 0.237548828125, 0.07177734375, 0.9326171875, -0.367919921875, -0.375732421875, -0.2314453125, -0.58349609375, -0.15185546875, 0.0180206298828125, -0.277099609375, 0.197021484375, 0.5791015625, 0.20068359375, 0.000244140625, 0.418701171875, -0.21923828125, 0.2529296875, -0.00341796875, -0.12060546875, 0.2291259765625, 0.281982421875, 0.19189453125, -0.222900390625, -0.337890625, -0.4853515625, -0.052734375, 0.245361328125, 0.0499267578125, 0.94140625, -0.0589599609375, -0.002197265625, -0.0888671875, -0.458740234375, 0.85693359375, -0.259765625, -0.45849609375, 0.21240234375, -0.712890625, 1.232421875, 0.273681640625, 0.54638671875, 0.4951171875, -0.323974609375, 0.03564453125, 0.145263671875, -0.04339599609375, 0.140625, -0.576171875, -0.17919921875, -0.1337890625, 0.0760498046875, -0.109619140625, 0.292236328125, 0.390380859375, 0.079833984375, 0.544921875, 0.0163421630859375, -0.63818359375, 0.7373046875, 0.355224609375, -0.37451171875, -0.6689453125, 0.0181884765625, -0.1053466796875, 0.03057861328125, -0.1002197265625, -0.1572265625, 0.020263671875, -0.06378173828125, 0.1865234375, 0.09881591796875, -0.472900390625, 0.14306640625, -0.6416015625, -0.2568359375, -0.177978515625, 0.373291015625, 0.5166015625, 0.0616455078125, -0.26953125, 0.00244140625, -0.71142578125, -0.0625, 0.2392578125, -0.0980224609375, 0.386962890625, 0.18310546875, 0.31591796875, -0.0501708984375, -0.37646484375, 0.04559326171875, 0.0306396484375, 0.591796875, -0.0628662109375, 0.049072265625, 0.78125, 0.487548828125, 0.023681640625, 0.0699462890625, 0.285400390625, 0.179443359375, -0.0673828125, -0.2333984375, 0.080078125, -0.012451171875, -0.321533203125, 0.203369140625, 1.068359375, -0.137451171875, -1.5830078125, 0.0528564453125, -0.3818359375, 0.391357421875, -0.2626953125, -0.2396240234375, 0.92333984375, 0.25439453125, -0.4140625, -0.29052734375, 0.28564453125, 0.4677734375, 0.333740234375, -0.3193359375, -0.2890625, 0.0218505859375, 0.6240234375, 0.0020751953125, 0.0673828125, 0.1539306640625, 0.399169921875, -0.193115234375, 0.63330078125, -0.24072265625, 0.99072265625, -1.173828125, 0.17236328125, -0.054931640625, -0.095947265625, 0.047607421875, 0.196533203125, -0.188232421875, 0.047515869140625, 0.0908203125, 0.00048828125, 0.04638671875, 0.21044921875, -0.058349609375, 0.175537109375, 0.225830078125, 0.132568359375, 0.476806640625, -0.38427734375, 0.53125, -0.1939697265625, 0.42626953125, -0.217041015625, -0.301513671875, 1.16015625, 0.461669921875, -0.033203125, -0.206787109375, 0.080078125, 0.01544189453125, 0.044097900390625, 0.67919921875, 0.05474853515625, -0.69287109375, -0.125244140625, 0.217529296875, -0.10986328125, 0.212158203125, -0.84326171875, -0.50830078125, -0.52392578125, 0.51171875, 1.048828125, -0.01262664794921875, -0.442626953125, -0.40576171875, -0.509765625, -0.1048583984375, -0.583984375, -0.1962890625, 0.241943359375, -0.1739501953125, -0.26611328125, 0.79638671875, 0.033447265625, -0.13037109375, 0.25048828125, 0.055908203125, 0.27001953125, -0.279541015625, -0.529296875, 0.73828125, -0.07568359375, -1.0390625, -0.0830078125, 0.71826171875, 0.259765625, -0.01788330078125, 0.10693359375, -0.19091796875, 0.93359375, -0.26513671875, 0.0240478515625, -0.35791015625, -0.188232421875, -0.039306640625, -0.04132080078125, 0.2802734375, -0.06201171875, -0.677734375, -0.095703125, -0.160888671875, -0.0498046875, 0.424560546875, 0.478515625, 0.00384521484375, -0.2139892578125, -0.067138671875, 0.0079345703125, -0.254638671875, 0.04736328125, -0.7734375, -0.326416015625, 0.6083984375, -0.43359375, 0.0694580078125, 0.392578125, 0.306396484375, 0.33203125, -0.263671875, -0.69580078125, -0.135498046875, 0.150146484375, -0.332275390625, -0.369140625, -0.256103515625, 0.6494140625, -0.45654296875, 0.125244140625, 0.417236328125, 0.313232421875, -0.13427734375, 0.5087890625, -0.55810546875, 1.154296875, -0.012603759765625, 0.047454833984375, 0.071533203125, -0.411376953125, 0.0377197265625, 0.126220703125, 0.202880859375, 0.40087890625, 0.623046875, -0.239013671875, 0.028564453125, 0.74365234375, 0.048065185546875, -0.042724609375, -0.3212890625, 0.39013671875, 0.281005859375, -0.68505859375, 0.10992431640625, -0.1270751953125, 0.354248046875, -0.43017578125, -0.00537109375, -0.33544921875, -0.36474609375, 0.42919921875, -0.11016845703125, 0.026123046875, -0.29052734375, 0.240966796875, -0.89404296875, 0.09423828125, 0.05682373046875, -0.4326171875, -0.857421875, 0.59716796875, -0.482421875, 0.337158203125, -0.541015625, -0.03204345703125, -0.46142578125, -0.124755859375, -0.19140625, -0.453369140625, 0.1729736328125, 0.201416015625, 0.14599609375, -0.6455078125, 0.02447509765625, -0.0281982421875, -0.72265625, 0.0218505859375, -0.310546875, -0.78076171875, -0.51953125, 0.137939453125, -0.442138671875, 0.283203125, 0.2039794921875, 0.0517578125, -0.3125, -0.404296875, 0.2003173828125, 0.134033203125, 0.42578125, 0.32958984375, 0.2393798828125, -0.1444091796875, -0.3505859375, -0.1807861328125, -0.186767578125, 0.0186767578125, 0.47021484375, 0.2919921875, -0.6484375, -0.46533203125, -0.04931640625, 0.287109375, 0.061065673828125, -0.069091796875, -0.02703857421875, 0.2509765625, -0.206787109375, 0.31640625, 0.29541015625, -0.0430908203125, 0.414306640625, -0.26708984375, 0.1630859375, -0.51953125, -0.0445556640625, -0.4150390625, 0.1474609375, 0.71630859375, 0.1895751953125, -0.091064453125, -0.04156494140625, -0.123291015625, -0.339599609375, 0.50341796875, -0.62890625, -0.2158203125, 0.2998046875, 0.2281494140625, 1.048828125, 0.6962890625, -0.143798828125, -0.6552734375, -0.06787109375, -0.3837890625, 0.228759765625, 0.149658203125, -0.121826171875, -0.44677734375, -0.14404296875, 0.02978515625, -0.340576171875, 0.07293701171875, 0.363525390625, -0.0279541015625, 0.214111328125, -0.9375, -0.7607421875, -0.275634765625, -0.3857421875, 0.5634765625, 0.736328125, -0.299560546875, 1.1171875, -0.252197265625, 0.17626953125, -0.0728759765625, -0.184326171875, -0.1029052734375, -0.0882568359375, -0.48681640625, -1.380859375, -0.0718994140625, 0.1260986328125, 0.212158203125, 0.1116943359375, -6.08203125, -0.447265625, -0.081298828125, -0.7353515625, -0.2039794921875, -0.0711669921875, -1.2197265625, -0.62939453125, -0.20849609375, -0.0228271484375, 0.51904296875, 0.64404296875, -0.249755859375, 0.091552734375, 0.2177734375, 0.283447265625, 0.03369140625, -0.0103759765625, 0.1990966796875, 0.0733642578125, 0.2158203125, -0.1495361328125, 0.0809326171875, 0.47607421875, -0.2841796875, -0.260986328125, 0.3876953125, -0.6875, -0.6630859375, -0.1025390625, 0.07568359375, -0.307861328125, -0.340087890625, -0.42138671875, -0.0657958984375, 0.578125, 0.7880859375, -0.3642578125, -0.34326171875, 0.1925048828125, -1.0126953125, 0.335693359375, -0.050048828125, -0.01934814453125, 0.247314453125, 0.378662109375, -0.08050537109375, -0.73046875, -0.755859375, 0.1494140625, 0.17578125, -1.033203125, -0.063232421875, 0.35302734375, -0.029541015625, 0.2236328125, -0.04364013671875, 0.0938720703125, 0.181396484375, 0.1705322265625, 0.5712890625, 0.339599609375, 0.430419921875, -0.33251953125, 0.018798828125, 0.058837890625, 0.020263671875, 0.11181640625, 0.08642578125, -0.0521240234375, 0.05352783203125, 0.65966796875, -0.173095703125, -0.291015625, -0.104248046875, -0.10528564453125, 0.0657958984375, -0.222900390625, 0.086181640625, 0.2230224609375, 0.0435791015625, 0.1009521484375, 0.327392578125, -0.22119140625, -0.1510009765625, -0.1064453125, -0.349365234375, -0.615234375, -0.118408203125, 0.72509765625, 0.136962890625, -0.183837890625, 0.125244140625, 0.1693115234375, 1.275390625, 0.181884765625, 0.1328125, 0.145751953125, 0.059326171875, -0.404052734375, -0.4521484375, -0.056640625, -0.082275390625, 0.34130859375, 0.276611328125, 0.085693359375, 0.01171875, 0.375, -0.0347900390625, 0.8125, 0.1473388671875, 0.42529296875, 0.53515625, -0.1934814453125, 0.487548828125, -1.5283203125, -0.43505859375, -0.21142578125, 0.1041259765625, -0.07861328125, -0.32470703125, 0.796875, 0.293701171875, -0.114990234375, 0.094970703125, 0.2919921875, -0.09375, -0.235595703125, -0.9248046875, -0.427001953125, -0.2587890625, -0.46875, -0.9541015625, 0.277099609375, 0.289306640625, 0.513671875, -0.44140625, -0.0643310546875, 0.045654296875, -0.0582275390625, 0.376953125, -0.271484375, 0.09588623046875, 0.163330078125, 0.41748046875, -0.380126953125, 0.1695556640625, -0.3447265625, -0.515625, -0.346435546875, 0.184326171875, -0.135009765625, -0.1014404296875, 0.05426025390625, -0.175048828125, -0.013671875, 0.45556640625, -0.335693359375, 0.16650390625, 0.37158203125, 0.00598907470703125, -0.23681640625, 0.061431884765625, 0.1343994140625, 0.031982421875, -0.097900390625, -0.215087890625, -1.4990234375, -0.48388671875, 0.1854248046875, -0.0777587890625, -0.342041015625, -0.37158203125, -0.242431640625, -0.115234375, -0.07562255859375, -0.2421875, -0.2412109375, -0.8564453125, 0.189697265625, 0.354736328125, -0.23291015625, -0.2275390625, -0.1392822265625, -0.59375, 0.2265625, 0.270751953125, -0.0521240234375, -0.2298583984375, -0.0108642578125, -0.0537109375, 0.00146484375, 0.0567626953125, -0.2666015625, -0.0112457275390625, 0.066650390625, 0.41796875, 0.291259765625, -0.0648193359375, 0.1409912109375, -0.0213623046875, -0.40087890625, 0.69677734375, 0.0455322265625, -0.148681640625, -0.30712890625, 1.341796875, 0.049072265625, -0.333984375, 0.0068359375, -0.418701171875, 0.338134765625, -0.260986328125, -0.21435546875, 0.32958984375, 0.1435546875, -0.035919189453125, 0.260498046875, 0.1385498046875, 0.26171875, -0.294189453125, 0.408203125, 0.20849609375, -0.59912109375, 0.18017578125, 0.31103515625, -0.09808349609375, -0.23095703125, -0.2381591796875, -0.295654296875, -0.8916015625, 0.0059814453125, -0.137451171875, -0.0870361328125, -0.31103515625, 0.08154296875, -0.12353515625, 0.457275390625, -0.460205078125, -1.419921875, -0.1002197265625, -0.1197509765625, 0.23974609375, 0.20458984375, -0.11737060546875, 0.289794921875, 0.1103515625, -0.148193359375, 0.294921875, 0.00079345703125, 0.1400146484375, 0.4609375, 0.02490234375, 0.1328125, -0.16748046875, 0.3212890625, 0.242431640625, -0.383056640625, -0.61279296875, -0.1953125, -0.439453125, 0.005126953125, 0.322265625, -0.246826171875, 0.841796875, -0.037353515625, 0.0970458984375, -0.84619140625, 0.010894775390625, -0.004638671875, 0.211181640625, -0.57080078125, 0.30322265625, -0.54296875, 0.4794921875, -0.02239990234375, 0.18310546875, -0.11883544921875, -0.5048828125, -0.76611328125, 0.095703125, -0.47900390625, 0.45263671875, -0.3603515625, 1.01171875, -0.34716796875, 0.086669921875, -0.093017578125, -0.063720703125, -0.4345703125, 0.09619140625, 1.44921875, 0.02294921875, 0.2464599609375, -0.06494140625, 0.412109375, -1.0478515625, -0.248291015625, 0.5390625, 0.078125, -0.2308349609375, 0.498046875, 0.1737060546875, 0.114013671875, 0.150390625, 0.03302001953125, 0.208984375, -0.0478515625, 0.439453125, -0.447265625, -0.017822265625, -0.67333984375, 0.058837890625, 0.09765625, -0.426025390625, -0.00433349609375, -0.147705078125, -0.1318359375, 0.106689453125, -0.8447265625, -0.27099609375, 0.2493896484375, -0.075927734375, 0.1767578125, -0.049896240234375, -0.112060546875, -0.7939453125, 0.0277099609375, 0.3046875, -0.08538818359375, 0.09735107421875, -0.20263671875, -0.499267578125, 0.5654296875, -0.41064453125, 0.01702880859375, -0.587890625, 0.05609130859375, -0.38232421875, 0.1121826171875, -0.0029296875, 0.0057373046875, 0.11297607421875, 0.22412109375, -2.3046875, 0.31982421875, -0.273193359375, 0.31689453125, 0.65185546875, -0.08349609375, -0.291015625, -0.26904296875, 0.16455078125, -0.1229248046875, 0.2734375, 0.045654296875, 0.0948486328125, 0.45166015625, 0.483154296875, 0.404052734375, -0.004150390625, 0.267822265625, 0.6767578125, -0.2939453125, 0.385009765625, 0.08349609375, 0.146240234375, -0.2413330078125, 0.1279296875, 0.291015625, 0.2978515625, -0.74365234375, 0.59375, 0.1317138671875, -0.0660400390625, 0.006103515625, -0.52880859375, -0.016845703125, -0.0006103515625, 0.292236328125, -0.32373046875, -0.0677490234375, -0.201416015625, -0.182373046875, 0.481689453125, -0.0386962890625, -0.28173828125, -0.6103515625, -0.399169921875, 0.242919921875, 0.1015625, -0.164794921875, 0.396240234375, -0.11016845703125, 0.269775390625, -0.414306640625, -0.27099609375, 0.1201171875, -0.314208984375, -0.505859375, -0.6689453125, -0.595703125, -0.0430908203125, 0.77294921875, 0.67724609375, -0.228271484375, 0.0350341796875, 0.6904296875, -0.6865234375, 0.31103515625, -0.7412109375, -0.5927734375, -0.50341796875, 0.08447265625, 0.0677490234375, -0.38623046875, -0.29443359375, 0.200439453125, -0.17431640625, 0.189453125, 0.429931640625, 0.0222625732421875, 0.182373046875, 0.0311279296875, -0.438720703125, 0.10205078125, -0.05670166015625, -0.104736328125, 0.78515625, 0.0224609375, -0.258544921875, 0.48046875, -0.6396484375, 0.452880859375, -0.271728515625, 0.599609375, -0.1778564453125, 0.072021484375, 0.1173095703125, -0.9609375, -0.0986328125, 0.0224609375, 0.149658203125, -0.28271484375, -0.3935546875, 0.0821533203125, 1.12890625, 0.462890625, -0.6259765625, -0.23974609375, 0.767578125, -0.172607421875, -0.1339111328125, -0.1016845703125, -0.42822265625, 0.262939453125, -0.171630859375, -0.46630859375, 0.44580078125, 0.4228515625, -0.3525390625, -0.5458984375, 0.12841796875, 0.074462890625, -0.093017578125, 0.1444091796875, -0.31787109375, -0.06036376953125, 0.1527099609375, 0.208251953125, 0.3359375, 0.3935546875, 0.24169921875, -0.1343994140625, 0.128662109375, -0.061767578125, 0.38818359375, 0.34912109375, 0.50390625, -0.248779296875, -0.171875, 0.138916015625, -0.264404296875, 0.00341796875, -0.26025390625, 0.3564453125, 0.43115234375, -0.278564453125, 0.438232421875, -2.146484375, 0.54541015625, 0.474609375, 0.10888671875, -0.5458984375, -0.160400390625, 0.0933837890625, -0.1611328125, 0.142822265625, 0.18505859375, 0.426513671875, -0.1611328125, 0.88037109375, -0.327392578125, -0.0152587890625, 0.62060546875, 0.2587890625, 0.26953125, -0.2236328125, -0.382080078125, 0.1981201171875, 0.4560546875, 0.15966796875, -0.02899169921875, -0.2164306640625, 0.2783203125, 0.18603515625, 0.97998046875, 0.26513671875, 0.331298828125, -0.14306640625, 0.3564453125, 0.155029296875, 0.02178955078125, 0.654296875, -0.7958984375, -0.360595703125, 0.54736328125, -0.5947265625, 0.0538330078125, 0.7978515625, -0.43310546875, -0.2298583984375, -0.0823974609375, -0.4150390625, -0.225341796875, 0.124755859375, -1.009765625, -0.04364013671875, 0.1292724609375, -0.016357421875, 0.625, -0.267578125, 0.39404296875, -0.330078125, -0.24267578125, -0.199462890625, -0.10205078125, 0.046142578125, -0.17236328125, 0.1239013671875, 0.0867919921875, 0.11578369140625, 0.24658203125, -0.20166015625, -0.156494140625, -0.36962890625, 0.035736083984375, 0.4892578125, 1.271484375, -0.180419921875, -0.291259765625, -0.5927734375, 0.552734375, 0.39111328125, 0.0693359375, 0.365234375, -0.27001953125, -1.2939453125, -0.0810546875, 0.010986328125, -0.6220703125, 0.39501953125, 0.5400390625, -0.0909423828125, 0.170654296875, -0.1513671875, 0.092041015625, 0.15966796875, 0.0516357421875, -0.4697265625, 0.2529296875, -0.138916015625, -0.436767578125, 0.1690673828125, -0.1285400390625, -0.3046875, -0.247314453125, -0.1883544921875, -0.1319580078125, -0.480224609375, 0.058349609375, -0.018798828125, -0.49365234375, -0.5078125, -0.152099609375, -0.45654296875, -0.016357421875, 0.2086181640625, -0.06298828125, -0.333251953125, -0.224365234375, -0.135986328125, -0.2252197265625, 0.261474609375, 0.494140625, 0.1527099609375, 0.2393798828125, -0.77099609375, -0.363037109375, -0.5166015625, -0.2235107421875, 0.02099609375, 0.173828125, 0.1942138671875, 0.1019287109375, 0.437744140625, -0.07122802734375, 0.157958984375, -0.040771484375, -0.1392822265625, 0.41015625, 0.261474609375, 0.60595703125, -0.1644287109375, 0.06298828125, -0.77587890625, -0.1094970703125, 0.06005859375, 0.2188720703125, -0.2568359375, -0.09271240234375, -0.5556640625, 0.27099609375, -1.408203125, 0.1243896484375, -0.09637451171875, 1.25390625, -0.1826171875, 0.3974609375, 0.3662109375, 0.441650390625, -0.1007080078125, -0.028472900390625, 0.298583984375, -0.0877685546875, 0.373046875, 0.60498046875, -0.64990234375, -0.22705078125, -0.1663818359375, -0.185546875, -0.09124755859375, 0.1533203125, 0.5078125, 0.0243072509765625, -0.58984375, 0.177490234375, -0.2041015625, -0.019378662109375, -0.7978515625, 0.85009765625, 0.009033203125, -0.0675048828125, 0.004150390625, -0.76806640625, 0.453369140625, 0.1878662109375, 0.5166015625, -0.01318359375, -0.6748046875, 0.6787109375, 0.2734375, -0.112548828125, 1.0966796875, 0.0616455078125, -0.036865234375, -0.6943359375, -0.2318115234375, -0.22119140625, 0.0244140625, 0.0869140625, 0.08404541015625, 0.47509765625, -0.177001953125, 0.1494140625, -0.0015869140625, -0.4970703125, -0.339111328125, 0.2939453125, -0.4140625, -0.26806640625, 0.1153564453125, -0.399658203125, -0.1336669921875, -0.06524658203125, -0.0810546875, 0.87353515625, 0.0074462890625, -0.275390625, -0.147705078125, 0.175048828125, 0.1328125, -0.0335693359375, 0.307373046875, -0.14208984375, 0.223388671875, 0.60693359375, -0.4716796875, 0.19580078125, -0.27099609375, -0.86767578125, 0.2119140625, 0.2626953125, 0.09375, -0.04656982421875, 0.05908203125, 0.134765625, 0.10693359375, -0.220458984375, 0.425537109375, 0.7373046875]}, {"frame_number": 15, "vector": [-0.1571044921875, 0.25537109375, 0.01953125, -0.1636962890625, -0.6328125, 0.30029296875, 0.015380859375, -0.472900390625, -0.10205078125, -4.57763671875e-05, -0.18359375, 0.07470703125, 0.31884765625, 0.1962890625, 0.01806640625, 0.286376953125, 0.1285400390625, -0.2109375, -0.1192626953125, -0.47216796875, -0.54541015625, -6.04296875, -0.368408203125, -0.76611328125, 0.039794921875, -0.013336181640625, -0.18505859375, -0.8251953125, 0.513671875, -0.215576171875, -0.9111328125, -0.5791015625, -0.63427734375, -0.054046630859375, -0.11859130859375, -0.010955810546875, 0.3017578125, -0.65234375, -0.08203125, -0.36376953125, 0.0352783203125, 0.1845703125, 0.222900390625, 0.05810546875, 0.9404296875, -0.408203125, -0.37646484375, -0.216552734375, -0.5947265625, -0.1435546875, 0.00844573974609375, -0.281982421875, 0.21826171875, 0.5908203125, 0.2060546875, -0.0120849609375, 0.388671875, -0.2357177734375, 0.239013671875, -0.0302734375, -0.1123046875, 0.247314453125, 0.281005859375, 0.208740234375, -0.188720703125, -0.32275390625, -0.50927734375, -0.049560546875, 0.253173828125, 0.018310546875, 0.9580078125, -0.0277099609375, 0.031005859375, -0.1064453125, -0.461181640625, 0.87744140625, -0.2421875, -0.453125, 0.27197265625, -0.71435546875, 1.24609375, 0.273681640625, 0.5576171875, 0.4873046875, -0.3056640625, 0.0408935546875, 0.11724853515625, -0.048980712890625, 0.16259765625, -0.5615234375, -0.1732177734375, -0.131103515625, 0.0718994140625, -0.07861328125, 0.29296875, 0.409423828125, 0.091796875, 0.53369140625, 0.0382080078125, -0.625, 0.72265625, 0.40576171875, -0.39794921875, -0.662109375, 0.04095458984375, -0.07763671875, 0.027252197265625, -0.06597900390625, -0.1566162109375, 0.051025390625, -0.06201171875, 0.189697265625, 0.09686279296875, -0.452392578125, 0.1304931640625, -0.6552734375, -0.24853515625, -0.1807861328125, 0.392333984375, 0.5078125, 0.0750732421875, -0.280029296875, 0.001953125, -0.67138671875, -0.040283203125, 0.271484375, -0.0914306640625, 0.37109375, 0.1573486328125, 0.29296875, -0.048828125, -0.361572265625, 0.0185546875, -0.0001220703125, 0.59716796875, -0.046142578125, 0.06964111328125, 0.77685546875, 0.490234375, -0.00732421875, 0.02099609375, 0.275634765625, 0.1881103515625, -0.0699462890625, -0.2236328125, 0.10595703125, -0.020294189453125, -0.330322265625, 0.203125, 1.0546875, -0.16650390625, -1.58984375, 0.019287109375, -0.3583984375, 0.369873046875, -0.2401123046875, -0.212646484375, 0.9208984375, 0.241943359375, -0.4423828125, -0.27587890625, 0.263427734375, 0.448486328125, 0.30908203125, -0.32470703125, -0.2568359375, -0.0008544921875, 0.57421875, -0.0269775390625, 0.0478515625, 0.190185546875, 0.37109375, -0.1724853515625, 0.65625, -0.215087890625, 0.96728515625, -1.197265625, 0.1982421875, -0.0723876953125, -0.09912109375, 0.0135498046875, 0.19189453125, -0.203857421875, 0.040557861328125, 0.0726318359375, 0.01397705078125, 0.080322265625, 0.220458984375, -0.0614013671875, 0.16845703125, 0.18212890625, 0.1151123046875, 0.469970703125, -0.39990234375, 0.53515625, -0.2249755859375, 0.444091796875, -0.1708984375, -0.238525390625, 1.162109375, 0.4462890625, -0.0224609375, -0.217041015625, 0.06396484375, 0.08306884765625, 0.039337158203125, 0.68505859375, 0.071044921875, -0.7568359375, -0.1533203125, 0.1883544921875, -0.112060546875, 0.20654296875, -0.85693359375, -0.49560546875, -0.53466796875, 0.52880859375, 1.00390625, -0.0155792236328125, -0.4453125, -0.4248046875, -0.48486328125, -0.0948486328125, -0.5634765625, -0.22119140625, 0.26708984375, -0.1461181640625, -0.2802734375, 0.77294921875, 0.03741455078125, -0.1708984375, 0.226318359375, 0.055419921875, 0.2685546875, -0.307373046875, -0.52490234375, 0.7333984375, -0.10125732421875, -1.060546875, -0.07958984375, 0.7392578125, 0.261962890625, -0.0814208984375, 0.11767578125, -0.193359375, 0.927734375, -0.29052734375, 0.0135498046875, -0.363525390625, -0.2138671875, -0.025146484375, -0.0390625, 0.28125, -0.068359375, -0.6640625, -0.083251953125, -0.151123046875, -0.054931640625, 0.420654296875, 0.45947265625, -0.00604248046875, -0.20751953125, -0.058349609375, 0.00390625, -0.20947265625, 0.028076171875, -0.77978515625, -0.34423828125, 0.6064453125, -0.40478515625, 0.06439208984375, 0.3486328125, 0.3115234375, 0.3330078125, -0.275390625, -0.6865234375, -0.151123046875, 0.1785888671875, -0.326416015625, -0.3701171875, -0.26318359375, 0.6689453125, -0.47216796875, 0.1015625, 0.385986328125, 0.30712890625, -0.131591796875, 0.5205078125, -0.544921875, 1.1708984375, -0.030242919921875, 0.0430908203125, 0.1014404296875, -0.39892578125, 0.033203125, 0.1246337890625, 0.216796875, 0.406494140625, 0.57861328125, -0.209228515625, 0.0164794921875, 0.75390625, 0.052764892578125, -0.020751953125, -0.323486328125, 0.37646484375, 0.28955078125, -0.673828125, 0.10791015625, -0.160888671875, 0.35400390625, -0.471923828125, 0.007568359375, -0.323486328125, -0.332275390625, 0.4111328125, -0.08148193359375, 0.026611328125, -0.289794921875, 0.249267578125, -0.89892578125, 0.1170654296875, 0.027099609375, -0.42724609375, -0.849609375, 0.60205078125, -0.4990234375, 0.33349609375, -0.53466796875, -0.0264892578125, -0.43798828125, -0.1263427734375, -0.202880859375, -0.43603515625, 0.1563720703125, 0.189453125, 0.16650390625, -0.6689453125, 0.01776123046875, -0.0296630859375, -0.712890625, 0.0301513671875, -0.322021484375, -0.78271484375, -0.533203125, 0.14111328125, -0.466552734375, 0.25341796875, 0.22021484375, 0.053466796875, -0.32373046875, -0.44677734375, 0.189697265625, 0.133544921875, 0.421630859375, 0.3427734375, 0.234619140625, -0.128173828125, -0.375, -0.1756591796875, -0.195556640625, 0.0228271484375, 0.49560546875, 0.292236328125, -0.6494140625, -0.474365234375, -0.08349609375, 0.281005859375, 0.0601806640625, -0.1033935546875, -0.01629638671875, 0.25341796875, -0.1982421875, 0.31591796875, 0.322265625, -0.0430908203125, 0.4365234375, -0.273681640625, 0.1865234375, -0.5419921875, -0.0498046875, -0.359375, 0.147216796875, 0.7060546875, 0.1865234375, -0.10888671875, -0.0228271484375, -0.128662109375, -0.307373046875, 0.5341796875, -0.63525390625, -0.24072265625, 0.2890625, 0.2255859375, 1.0673828125, 0.68896484375, -0.177734375, -0.638671875, -0.083740234375, -0.399169921875, 0.239013671875, 0.1453857421875, -0.143310546875, -0.47021484375, -0.128173828125, 0.0465087890625, -0.319091796875, 0.08935546875, 0.37158203125, -0.0404052734375, 0.239501953125, -0.90966796875, -0.73095703125, -0.28857421875, -0.41552734375, 0.5517578125, 0.6982421875, -0.30859375, 1.138671875, -0.194580078125, 0.1666259765625, -0.0386962890625, -0.15771484375, -0.124755859375, -0.09130859375, -0.4931640625, -1.4033203125, -0.056884765625, 0.1314697265625, 0.228759765625, 0.139892578125, -6.04296875, -0.4111328125, -0.063232421875, -0.74365234375, -0.1934814453125, -0.0860595703125, -1.21484375, -0.61865234375, -0.20361328125, -0.02520751953125, 0.541015625, 0.6240234375, -0.218994140625, 0.1005859375, 0.2314453125, 0.262451171875, 0.0224609375, -0.015380859375, 0.21533203125, 0.08251953125, 0.22412109375, -0.1402587890625, 0.095947265625, 0.4736328125, -0.2958984375, -0.27001953125, 0.382568359375, -0.7041015625, -0.65380859375, -0.08056640625, 0.0625, -0.32958984375, -0.345703125, -0.421875, -0.06298828125, 0.57763671875, 0.7900390625, -0.3310546875, -0.376953125, 0.1929931640625, -0.98193359375, 0.31787109375, -0.0775146484375, -0.03363037109375, 0.218994140625, 0.4150390625, -0.0711669921875, -0.763671875, -0.7607421875, 0.170166015625, 0.1741943359375, -1.0302734375, -0.060791015625, 0.354248046875, -0.0303955078125, 0.23046875, -0.02947998046875, 0.1109619140625, 0.1842041015625, 0.147216796875, 0.5439453125, 0.3134765625, 0.433837890625, -0.319580078125, 0.000244140625, 0.03985595703125, -6.103515625e-05, 0.1162109375, 0.0838623046875, -0.0582275390625, 0.06170654296875, 0.70556640625, -0.1937255859375, -0.30322265625, -0.0966796875, -0.1119384765625, 0.0244140625, -0.21630859375, 0.096435546875, 0.1988525390625, 0.033447265625, 0.0771484375, 0.30615234375, -0.242431640625, -0.1878662109375, -0.06005859375, -0.37451171875, -0.60546875, -0.104248046875, 0.71875, 0.15087890625, -0.150146484375, 0.138671875, 0.1771240234375, 1.2861328125, 0.17919921875, 0.0968017578125, 0.1365966796875, 0.05029296875, -0.44775390625, -0.467041015625, -0.05810546875, -0.083251953125, 0.296875, 0.2548828125, 0.10986328125, 0.0194091796875, 0.37548828125, -0.028564453125, 0.82373046875, 0.1416015625, 0.407958984375, 0.560546875, -0.172607421875, 0.490234375, -1.48828125, -0.457275390625, -0.177490234375, 0.1005859375, -0.08154296875, -0.3486328125, 0.8017578125, 0.294921875, -0.1290283203125, 0.11376953125, 0.309814453125, -0.11572265625, -0.222900390625, -0.8994140625, -0.41552734375, -0.258056640625, -0.4580078125, -0.97265625, 0.29736328125, 0.298095703125, 0.5107421875, -0.43115234375, -0.05889892578125, 0.0552978515625, -0.0950927734375, 0.37109375, -0.263671875, 0.0814208984375, 0.157958984375, 0.45166015625, -0.4140625, 0.1651611328125, -0.349365234375, -0.51513671875, -0.3505859375, 0.21533203125, -0.12744140625, -0.078369140625, 0.084228515625, -0.1917724609375, 0.01171875, 0.43017578125, -0.319091796875, 0.1741943359375, 0.363525390625, 0.003936767578125, -0.234619140625, 0.05560302734375, 0.10687255859375, 0.03192138671875, -0.099853515625, -0.21728515625, -1.521484375, -0.45361328125, 0.1943359375, -0.04766845703125, -0.3515625, -0.363525390625, -0.26806640625, -0.113037109375, -0.061553955078125, -0.2222900390625, -0.2421875, -0.86376953125, 0.1685791015625, 0.33837890625, -0.2327880859375, -0.214111328125, -0.170166015625, -0.61328125, 0.2286376953125, 0.27490234375, -0.053375244140625, -0.222900390625, -0.01123046875, -0.044677734375, 0.003173828125, 0.0711669921875, -0.28173828125, -0.0158538818359375, 0.03985595703125, 0.44140625, 0.2763671875, -0.06103515625, 0.1627197265625, -0.040252685546875, -0.384033203125, 0.6689453125, 0.02886962890625, -0.15185546875, -0.2734375, 1.3388671875, 0.043212890625, -0.356689453125, -0.0068359375, -0.36572265625, 0.3369140625, -0.25146484375, -0.212890625, 0.348388671875, 0.1724853515625, -0.00360107421875, 0.26708984375, 0.131591796875, 0.2685546875, -0.30810546875, 0.41552734375, 0.2265625, -0.6240234375, 0.17529296875, 0.3125, -0.0885009765625, -0.2314453125, -0.2225341796875, -0.328125, -0.90966796875, -0.0120849609375, -0.12298583984375, -0.12158203125, -0.33837890625, 0.094970703125, -0.153076171875, 0.43994140625, -0.46484375, -1.4140625, -0.10595703125, -0.1142578125, 0.2333984375, 0.1917724609375, -0.1536865234375, 0.24755859375, 0.125732421875, -0.14306640625, 0.338623046875, 0.01153564453125, 0.1251220703125, 0.444091796875, 0.029296875, 0.1082763671875, -0.126708984375, 0.32763671875, 0.23095703125, -0.3583984375, -0.646484375, -0.170654296875, -0.423095703125, 0.01806640625, 0.337890625, -0.255615234375, 0.810546875, -0.0400390625, 0.11328125, -0.82568359375, 0.0389404296875, 0.0128173828125, 0.22705078125, -0.5654296875, 0.31005859375, -0.578125, 0.50244140625, -0.02239990234375, 0.176513671875, -0.0966796875, -0.53759765625, -0.76611328125, 0.08319091796875, -0.4677734375, 0.44873046875, -0.312255859375, 1.03125, -0.341064453125, 0.096435546875, -0.115234375, -0.06866455078125, -0.44970703125, 0.106689453125, 1.443359375, 0.038238525390625, 0.215576171875, -0.0362548828125, 0.416015625, -1.033203125, -0.224365234375, 0.5517578125, 0.0780029296875, -0.263916015625, 0.484375, 0.1588134765625, 0.11865234375, 0.1376953125, 0.037353515625, 0.191162109375, -0.037841796875, 0.428466796875, -0.41845703125, -0.033203125, -0.677734375, 0.0943603515625, 0.093505859375, -0.45947265625, -0.01031494140625, -0.13427734375, -0.162841796875, 0.123046875, -0.83056640625, -0.27294921875, 0.25830078125, -0.0743408203125, 0.1807861328125, -0.050750732421875, -0.12158203125, -0.77734375, 0.001220703125, 0.27001953125, -0.10107421875, 0.093994140625, -0.20068359375, -0.5107421875, 0.5390625, -0.4150390625, 0.0504150390625, -0.58642578125, 0.04339599609375, -0.34716796875, 0.078369140625, 0.02978515625, -0.01611328125, 0.12249755859375, 0.260009765625, -2.326171875, 0.314208984375, -0.289306640625, 0.33984375, 0.654296875, -0.0828857421875, -0.31640625, -0.25, 0.1416015625, -0.1280517578125, 0.303466796875, 0.07177734375, 0.1187744140625, 0.426025390625, 0.468017578125, 0.415771484375, -0.0040283203125, 0.275390625, 0.69580078125, -0.29248046875, 0.38720703125, 0.0989990234375, 0.1588134765625, -0.218505859375, 0.09033203125, 0.307373046875, 0.33837890625, -0.75, 0.59423828125, 0.10498046875, -0.05450439453125, -0.0003662109375, -0.5205078125, -0.0247802734375, 0.039947509765625, 0.299560546875, -0.33251953125, -0.0545654296875, -0.203125, -0.174072265625, 0.51318359375, -0.022796630859375, -0.303955078125, -0.6201171875, -0.376220703125, 0.25146484375, 0.0802001953125, -0.1722412109375, 0.38232421875, -0.10662841796875, 0.2626953125, -0.40283203125, -0.270751953125, 0.1185302734375, -0.31298828125, -0.46630859375, -0.6328125, -0.59521484375, -0.043701171875, 0.76806640625, 0.67822265625, -0.26513671875, 0.0355224609375, 0.6796875, -0.71142578125, 0.329833984375, -0.7578125, -0.5693359375, -0.48974609375, 0.05126953125, 0.06536865234375, -0.39453125, -0.2822265625, 0.188720703125, -0.1884765625, 0.17822265625, 0.43359375, 0.032806396484375, 0.200927734375, 0.02899169921875, -0.4677734375, 0.108154296875, -0.0634765625, -0.10107421875, 0.81591796875, 0.005615234375, -0.25390625, 0.42626953125, -0.64599609375, 0.44384765625, -0.217529296875, 0.56689453125, -0.1617431640625, 0.06787109375, 0.1439208984375, -0.9853515625, -0.099609375, 0.0235595703125, 0.136962890625, -0.28662109375, -0.37744140625, 0.0780029296875, 1.1015625, 0.449951171875, -0.611328125, -0.22705078125, 0.73681640625, -0.182373046875, -0.135986328125, -0.1005859375, -0.4130859375, 0.24169921875, -0.1856689453125, -0.4931640625, 0.4931640625, 0.41259765625, -0.3701171875, -0.57080078125, 0.1292724609375, 0.1173095703125, -0.0772705078125, 0.142822265625, -0.317138671875, -0.0755615234375, 0.1448974609375, 0.1612548828125, 0.34423828125, 0.3818359375, 0.2490234375, -0.1605224609375, 0.13427734375, -0.064208984375, 0.360595703125, 0.354736328125, 0.475830078125, -0.2578125, -0.2088623046875, 0.15869140625, -0.241943359375, 0.0126953125, -0.28125, 0.355224609375, 0.44970703125, -0.28369140625, 0.419189453125, -2.15234375, 0.552734375, 0.4833984375, 0.098876953125, -0.552734375, -0.1456298828125, 0.083984375, -0.168701171875, 0.1365966796875, 0.1728515625, 0.4375, -0.162841796875, 0.873046875, -0.309814453125, -0.0074462890625, 0.59619140625, 0.251953125, 0.244140625, -0.2119140625, -0.390625, 0.21923828125, 0.46240234375, 0.1650390625, 0.00384521484375, -0.2724609375, 0.2939453125, 0.18017578125, 0.98583984375, 0.26171875, 0.339599609375, -0.113525390625, 0.372802734375, 0.13916015625, 0.0489501953125, 0.6669921875, -0.78125, -0.35302734375, 0.5146484375, -0.57470703125, 0.0401611328125, 0.8154296875, -0.4208984375, -0.2484130859375, -0.079345703125, -0.443359375, -0.2080078125, 0.1180419921875, -1.0126953125, -0.0230712890625, 0.10504150390625, -0.0159912109375, 0.62646484375, -0.27734375, 0.380859375, -0.371826171875, -0.24462890625, -0.203125, -0.134033203125, 0.039306640625, -0.204833984375, 0.133544921875, 0.06793212890625, 0.0947265625, 0.220947265625, -0.199462890625, -0.155029296875, -0.387939453125, 0.0176239013671875, 0.51806640625, 1.2607421875, -0.19287109375, -0.3203125, -0.55859375, 0.5302734375, 0.38916015625, 0.09375, 0.2919921875, -0.27587890625, -1.2802734375, -0.0615234375, 0.0196533203125, -0.6220703125, 0.40673828125, 0.537109375, -0.103515625, 0.161376953125, -0.15625, 0.100341796875, 0.1923828125, 0.0638427734375, -0.474609375, 0.2626953125, -0.1181640625, -0.41748046875, 0.2015380859375, -0.1197509765625, -0.29345703125, -0.271240234375, -0.20654296875, -0.0911865234375, -0.471435546875, 0.073974609375, 0.0057373046875, -0.509765625, -0.486328125, -0.173095703125, -0.47021484375, -0.0050048828125, 0.2083740234375, -0.060791015625, -0.3232421875, -0.236083984375, -0.16650390625, -0.236572265625, 0.243896484375, 0.46337890625, 0.1553955078125, 0.23193359375, -0.78515625, -0.351806640625, -0.54052734375, -0.2344970703125, 0.0068359375, 0.1783447265625, 0.1873779296875, 0.0772705078125, 0.433837890625, -0.0565185546875, 0.1873779296875, -0.046142578125, -0.1588134765625, 0.41943359375, 0.247802734375, 0.60400390625, -0.200439453125, 0.06494140625, -0.7802734375, -0.12158203125, 0.0552978515625, 0.2132568359375, -0.263427734375, -0.1058349609375, -0.53125, 0.276123046875, -1.392578125, 0.1270751953125, -0.09783935546875, 1.2470703125, -0.1949462890625, 0.40234375, 0.36279296875, 0.413330078125, -0.098388671875, -0.00390625, 0.30859375, -0.0968017578125, 0.372314453125, 0.599609375, -0.64013671875, -0.226318359375, -0.1962890625, -0.1907958984375, -0.1361083984375, 0.17138671875, 0.50390625, 0.0103759765625, -0.59765625, 0.188232421875, -0.206787109375, -0.05267333984375, -0.78515625, 0.8271484375, 0.0206298828125, -0.0338134765625, 0.013427734375, -0.7939453125, 0.4580078125, 0.1766357421875, 0.51953125, 0.002197265625, -0.6484375, 0.6513671875, 0.290283203125, -0.1290283203125, 1.115234375, 0.0308837890625, -0.0694580078125, -0.6640625, -0.2257080078125, -0.21826171875, 0.0167236328125, 0.072509765625, 0.066162109375, 0.48388671875, -0.16357421875, 0.12445068359375, 0.0250244140625, -0.52685546875, -0.322998046875, 0.295166015625, -0.400390625, -0.2274169921875, 0.114990234375, -0.429443359375, -0.1259765625, -0.047576904296875, -0.08984375, 0.876953125, -0.0169677734375, -0.259521484375, -0.14794921875, 0.1871337890625, 0.1533203125, -0.051513671875, 0.310791015625, -0.1536865234375, 0.2203369140625, 0.6337890625, -0.45751953125, 0.192138671875, -0.262451171875, -0.84130859375, 0.225830078125, 0.2412109375, 0.1044921875, -0.03271484375, 0.04736328125, 0.134765625, 0.1002197265625, -0.209716796875, 0.4091796875, 0.73486328125]}, {"frame_number": 16, "vector": [-0.1556396484375, 0.236572265625, -0.009521484375, -0.15283203125, -0.6318359375, 0.27880859375, -0.00836181640625, -0.48291015625, -0.1177978515625, -0.00262451171875, -0.166015625, 0.09228515625, 0.29296875, 0.204345703125, -0.006103515625, 0.292236328125, 0.139404296875, -0.21728515625, -0.10400390625, -0.50439453125, -0.53759765625, -5.97265625, -0.37744140625, -0.779296875, 0.04840087890625, -0.009185791015625, -0.16796875, -0.83203125, 0.49560546875, -0.21875, -0.931640625, -0.60595703125, -0.666015625, -0.06396484375, -0.1207275390625, 0.0225830078125, 0.3125, -0.66748046875, -0.07647705078125, -0.352294921875, 0.0101318359375, 0.205810546875, 0.2275390625, 0.021728515625, 0.9248046875, -0.406005859375, -0.39208984375, -0.2109375, -0.595703125, -0.22021484375, 0.022857666015625, -0.256591796875, 0.2276611328125, 0.6015625, 0.20703125, -0.03289794921875, 0.402587890625, -0.230224609375, 0.2366943359375, -0.01446533203125, -0.10986328125, 0.263916015625, 0.26025390625, 0.2183837890625, -0.1986083984375, -0.3134765625, -0.49755859375, -0.100830078125, 0.244384765625, -0.021240234375, 0.9677734375, -0.0224609375, 0.048095703125, -0.0908203125, -0.47509765625, 0.85498046875, -0.249267578125, -0.46337890625, 0.272216796875, -0.6904296875, 1.2734375, 0.25634765625, 0.56494140625, 0.501953125, -0.3037109375, 0.0386962890625, 0.109375, -0.063720703125, 0.17431640625, -0.552734375, -0.1978759765625, -0.13671875, 0.072021484375, -0.071044921875, 0.29638671875, 0.3955078125, 0.10498046875, 0.56201171875, 0.0153656005859375, -0.60986328125, 0.7197265625, 0.410888671875, -0.40869140625, -0.646484375, 0.055419921875, -0.1055908203125, 0.02154541015625, -0.051727294921875, -0.160400390625, -0.025146484375, -0.0472412109375, 0.186279296875, 0.086669921875, -0.46875, 0.1053466796875, -0.65625, -0.226318359375, -0.15234375, 0.37158203125, 0.513671875, 0.0692138671875, -0.286376953125, -0.0072021484375, -0.70166015625, -0.02435302734375, 0.288330078125, -0.088623046875, 0.369873046875, 0.169677734375, 0.31201171875, -0.0201416015625, -0.364990234375, 0.0343017578125, 0.014404296875, 0.6171875, -0.0501708984375, 0.05938720703125, 0.7705078125, 0.480224609375, -0.009521484375, 0.0283203125, 0.24365234375, 0.2044677734375, -0.07275390625, -0.21337890625, 0.093017578125, -0.001995086669921875, -0.34619140625, 0.2088623046875, 1.0400390625, -0.2095947265625, -1.587890625, 0.01416015625, -0.30908203125, 0.356689453125, -0.22021484375, -0.2149658203125, 0.923828125, 0.197265625, -0.458984375, -0.278564453125, 0.266845703125, 0.45068359375, 0.3076171875, -0.333984375, -0.263916015625, 0.003173828125, 0.59521484375, -0.020751953125, 0.017333984375, 0.203857421875, 0.388427734375, -0.1475830078125, 0.6552734375, -0.20751953125, 0.970703125, -1.20703125, 0.2166748046875, -0.034423828125, -0.1064453125, 0.009521484375, 0.14306640625, -0.2025146484375, 0.011138916015625, 0.07281494140625, 0.00537109375, 0.07470703125, 0.2353515625, -0.0838623046875, 0.130859375, 0.1922607421875, 0.11572265625, 0.4677734375, -0.37890625, 0.50927734375, -0.216064453125, 0.441162109375, -0.17529296875, -0.208984375, 1.16015625, 0.43359375, -0.0078125, -0.245849609375, 0.054443359375, 0.08392333984375, 0.0172119140625, 0.6484375, 0.07373046875, -0.779296875, -0.172119140625, 0.1995849609375, -0.093994140625, 0.2296142578125, -0.83740234375, -0.473388671875, -0.51611328125, 0.52783203125, 0.98681640625, -0.03399658203125, -0.476318359375, -0.40673828125, -0.525390625, -0.096923828125, -0.56787109375, -0.232666015625, 0.25830078125, -0.139892578125, -0.28857421875, 0.78759765625, 0.03973388671875, -0.1644287109375, 0.212646484375, 0.087646484375, 0.235107421875, -0.364990234375, -0.5185546875, 0.7421875, -0.148193359375, -1.0234375, -0.060791015625, 0.73681640625, 0.239990234375, -0.10772705078125, 0.121826171875, -0.2119140625, 0.908203125, -0.316162109375, 0.0018310546875, -0.35693359375, -0.220458984375, -0.053955078125, -0.04937744140625, 0.25537109375, -0.070068359375, -0.669921875, -0.0830078125, -0.1314697265625, -0.066650390625, 0.411376953125, 0.459716796875, 0.010009765625, -0.2135009765625, -0.079833984375, 0.010498046875, -0.2109375, 0.047119140625, -0.74609375, -0.32275390625, 0.59130859375, -0.4326171875, 0.0675048828125, 0.31884765625, 0.30810546875, 0.3203125, -0.28759765625, -0.68505859375, -0.1439208984375, 0.18017578125, -0.315185546875, -0.35693359375, -0.2357177734375, 0.6630859375, -0.468505859375, 0.1005859375, 0.39990234375, 0.29833984375, -0.1513671875, 0.47509765625, -0.509765625, 1.15625, -0.00408935546875, 0.01641845703125, 0.1378173828125, -0.430908203125, 0.0225830078125, 0.099853515625, 0.236572265625, 0.43896484375, 0.541015625, -0.220947265625, -0.0201416015625, 0.771484375, 0.056671142578125, -0.015380859375, -0.3037109375, 0.388916015625, 0.295166015625, -0.708984375, 0.114990234375, -0.1552734375, 0.348388671875, -0.49462890625, 0.00732421875, -0.307861328125, -0.309326171875, 0.427734375, -0.09332275390625, 0.0426025390625, -0.2666015625, 0.239013671875, -0.89404296875, 0.0916748046875, -0.00115966796875, -0.41259765625, -0.84765625, 0.611328125, -0.47998046875, 0.34912109375, -0.513671875, -0.013671875, -0.455078125, -0.1239013671875, -0.2061767578125, -0.421875, 0.15478515625, 0.221923828125, 0.16064453125, -0.66015625, -0.0076904296875, -0.0081787109375, -0.69384765625, 0.03826904296875, -0.32568359375, -0.81005859375, -0.5400390625, 0.11181640625, -0.46533203125, 0.239990234375, 0.201416015625, 0.058837890625, -0.3134765625, -0.44970703125, 0.1788330078125, 0.138671875, 0.407470703125, 0.28857421875, 0.24365234375, -0.161376953125, -0.3583984375, -0.171142578125, -0.1953125, -0.0037841796875, 0.5166015625, 0.28173828125, -0.62890625, -0.480712890625, -0.072509765625, 0.26513671875, 0.06854248046875, -0.093994140625, -0.01812744140625, 0.2476806640625, -0.18701171875, 0.293701171875, 0.30517578125, -0.01409912109375, 0.423828125, -0.259765625, 0.20703125, -0.54443359375, -0.0360107421875, -0.384765625, 0.145751953125, 0.7021484375, 0.1737060546875, -0.118408203125, 0.0029296875, -0.157958984375, -0.305908203125, 0.51953125, -0.62890625, -0.256591796875, 0.3046875, 0.22705078125, 1.060546875, 0.669921875, -0.20654296875, -0.625, -0.089111328125, -0.415283203125, 0.2265625, 0.11627197265625, -0.1484375, -0.43359375, -0.1259765625, 0.044189453125, -0.29833984375, 0.10107421875, 0.355224609375, -0.008544921875, 0.26416015625, -0.90087890625, -0.71875, -0.2822265625, -0.423095703125, 0.56640625, 0.71240234375, -0.3017578125, 1.1474609375, -0.1942138671875, 0.1802978515625, -0.025146484375, -0.135498046875, -0.1239013671875, -0.109130859375, -0.5283203125, -1.380859375, -0.07293701171875, 0.1241455078125, 0.2142333984375, 0.1484375, -5.984375, -0.3984375, -0.0341796875, -0.751953125, -0.221923828125, -0.0670166015625, -1.2109375, -0.5927734375, -0.19384765625, -0.01739501953125, 0.546875, 0.61865234375, -0.195068359375, 0.1038818359375, 0.21533203125, 0.28271484375, -0.002197265625, -0.014404296875, 0.240478515625, 0.09912109375, 0.214599609375, -0.17138671875, 0.0401611328125, 0.48291015625, -0.316650390625, -0.260009765625, 0.37646484375, -0.70654296875, -0.67626953125, -0.052001953125, 0.061279296875, -0.3203125, -0.345703125, -0.419921875, -0.032958984375, 0.5771484375, 0.77294921875, -0.33447265625, -0.3408203125, 0.178955078125, -0.9814453125, 0.31103515625, -0.075439453125, -0.0233154296875, 0.21728515625, 0.4599609375, -0.0689697265625, -0.77734375, -0.72607421875, 0.170166015625, 0.1875, -1.0244140625, -0.041748046875, 0.360107421875, -0.01312255859375, 0.263916015625, -0.02630615234375, 0.1258544921875, 0.171875, 0.1385498046875, 0.5283203125, 0.30908203125, 0.413330078125, -0.31103515625, -0.013916015625, 0.053985595703125, 0.015380859375, 0.1025390625, 0.09130859375, -0.0809326171875, 0.067138671875, 0.71337890625, -0.160400390625, -0.3056640625, -0.087646484375, -0.10125732421875, 0.054931640625, -0.2314453125, 0.09228515625, 0.213134765625, 0.0325927734375, 0.07208251953125, 0.316650390625, -0.241455078125, -0.178466796875, -0.061279296875, -0.383056640625, -0.6171875, -0.097900390625, 0.716796875, 0.150146484375, -0.1663818359375, 0.1295166015625, 0.1766357421875, 1.2822265625, 0.16357421875, 0.08880615234375, 0.1397705078125, 0.04541015625, -0.41162109375, -0.485107421875, -0.048828125, -0.108642578125, 0.2734375, 0.24267578125, 0.128662109375, 0.0111083984375, 0.40673828125, -0.0179443359375, 0.8017578125, 0.1343994140625, 0.414306640625, 0.54736328125, -0.163818359375, 0.50390625, -1.466796875, -0.45263671875, -0.17138671875, 0.0980224609375, -0.089111328125, -0.37646484375, 0.80712890625, 0.301513671875, -0.1217041015625, 0.126953125, 0.298095703125, -0.1134033203125, -0.187255859375, -0.8876953125, -0.418212890625, -0.245849609375, -0.465576171875, -0.99853515625, 0.322265625, 0.320556640625, 0.48388671875, -0.40576171875, -0.07476806640625, 0.0576171875, -0.084716796875, 0.346435546875, -0.25, 0.06512451171875, 0.159423828125, 0.462158203125, -0.404541015625, 0.1917724609375, -0.370361328125, -0.53271484375, -0.385498046875, 0.224609375, -0.142578125, -0.0712890625, 0.06756591796875, -0.2022705078125, -0.00830078125, 0.443359375, -0.29150390625, 0.1595458984375, 0.367919921875, -0.034210205078125, -0.2298583984375, 0.0904541015625, 0.10430908203125, 0.00927734375, -0.0970458984375, -0.201416015625, -1.529296875, -0.47412109375, 0.173828125, -0.0213623046875, -0.353515625, -0.351318359375, -0.292236328125, -0.096923828125, -0.046051025390625, -0.219482421875, -0.219482421875, -0.8544921875, 0.13720703125, 0.34619140625, -0.2298583984375, -0.174560546875, -0.1707763671875, -0.60205078125, 0.23974609375, 0.279541015625, -0.0699462890625, -0.20751953125, -0.0262451171875, -0.015380859375, 0.0125732421875, 0.0758056640625, -0.2587890625, -0.0068359375, 0.043121337890625, 0.447021484375, 0.28515625, -0.063720703125, 0.1773681640625, -0.07440185546875, -0.365234375, 0.67578125, 0.0074462890625, -0.16552734375, -0.28857421875, 1.3251953125, 0.066162109375, -0.33984375, -0.011474609375, -0.36767578125, 0.328125, -0.209228515625, -0.18896484375, 0.366943359375, 0.206787109375, -0.00469970703125, 0.251220703125, 0.132080078125, 0.28173828125, -0.32666015625, 0.41455078125, 0.23193359375, -0.6376953125, 0.182373046875, 0.29052734375, -0.08935546875, -0.23583984375, -0.18017578125, -0.344970703125, -0.837890625, 0.006103515625, -0.1378173828125, -0.1578369140625, -0.314208984375, 0.076904296875, -0.13427734375, 0.4208984375, -0.4384765625, -1.4072265625, -0.1328125, -0.0972900390625, 0.20458984375, 0.2060546875, -0.1346435546875, 0.2279052734375, 0.1243896484375, -0.14208984375, 0.350830078125, 0.01458740234375, 0.10595703125, 0.4345703125, 0.021728515625, 0.093017578125, -0.12890625, 0.34375, 0.239990234375, -0.331298828125, -0.6533203125, -0.1612548828125, -0.4267578125, 0.0244140625, 0.3544921875, -0.246826171875, 0.818359375, -0.0499267578125, 0.0972900390625, -0.7978515625, 0.0582275390625, 0.0291748046875, 0.22900390625, -0.5625, 0.324462890625, -0.6083984375, 0.4658203125, -0.009246826171875, 0.177978515625, -0.11328125, -0.5498046875, -0.78173828125, 0.082275390625, -0.4404296875, 0.43994140625, -0.296630859375, 1.01953125, -0.335205078125, 0.122314453125, -0.10693359375, -0.060302734375, -0.4619140625, 0.0927734375, 1.4453125, 0.022705078125, 0.2425537109375, -0.014892578125, 0.429931640625, -1.0048828125, -0.25390625, 0.52978515625, 0.104736328125, -0.2320556640625, 0.473388671875, 0.16943359375, 0.103271484375, 0.12548828125, 0.03094482421875, 0.202392578125, 0.0025634765625, 0.42724609375, -0.419189453125, 0.002197265625, -0.7041015625, 0.128662109375, 0.0960693359375, -0.4609375, -0.00311279296875, -0.13525390625, -0.16943359375, 0.111572265625, -0.84521484375, -0.326416015625, 0.27685546875, -0.07293701171875, 0.18896484375, -0.06732177734375, -0.135986328125, -0.7724609375, 0.0048828125, 0.274169921875, -0.07916259765625, 0.10223388671875, -0.218017578125, -0.51611328125, 0.52734375, -0.392822265625, 0.0313720703125, -0.60888671875, 0.04278564453125, -0.307861328125, 0.09283447265625, 0.06201171875, -0.056640625, 0.1220703125, 0.24560546875, -2.30859375, 0.333984375, -0.308349609375, 0.330078125, 0.67919921875, -0.0947265625, -0.34619140625, -0.255859375, 0.14501953125, -0.1131591796875, 0.307373046875, 0.08349609375, 0.1617431640625, 0.4326171875, 0.47119140625, 0.41943359375, 0.01025390625, 0.256591796875, 0.7021484375, -0.2900390625, 0.368896484375, 0.0819091796875, 0.1768798828125, -0.2041015625, 0.104248046875, 0.29296875, 0.34912109375, -0.7578125, 0.58837890625, 0.125, -0.0567626953125, -0.01904296875, -0.55810546875, -0.0244140625, 0.014007568359375, 0.281005859375, -0.308349609375, -0.052093505859375, -0.2080078125, -0.18994140625, 0.537109375, -0.0472412109375, -0.303955078125, -0.6201171875, -0.385986328125, 0.244873046875, 0.0821533203125, -0.1802978515625, 0.396240234375, -0.13232421875, 0.2822265625, -0.40283203125, -0.256591796875, 0.1551513671875, -0.27978515625, -0.4599609375, -0.62890625, -0.60546875, -0.04718017578125, 0.75634765625, 0.6845703125, -0.24560546875, 0.04388427734375, 0.6923828125, -0.6953125, 0.350341796875, -0.75244140625, -0.54736328125, -0.50146484375, 0.041015625, 0.0697021484375, -0.370849609375, -0.257080078125, 0.1751708984375, -0.193359375, 0.18701171875, 0.441162109375, 0.072509765625, 0.210693359375, 0.031585693359375, -0.466064453125, 0.109375, -0.04937744140625, -0.1033935546875, 0.802734375, 0.008056640625, -0.248046875, 0.43310546875, -0.69091796875, 0.475341796875, -0.227783203125, 0.56298828125, -0.14990234375, 0.03076171875, 0.15283203125, -0.9765625, -0.081787109375, 0.0335693359375, 0.14453125, -0.281494140625, -0.40478515625, 0.0830078125, 1.0986328125, 0.47509765625, -0.60693359375, -0.23193359375, 0.74658203125, -0.174072265625, -0.1387939453125, -0.106689453125, -0.39794921875, 0.259033203125, -0.1815185546875, -0.50341796875, 0.53125, 0.43212890625, -0.3740234375, -0.54296875, 0.1346435546875, 0.137939453125, -0.07489013671875, 0.1422119140625, -0.31103515625, -0.0830078125, 0.1405029296875, 0.1268310546875, 0.349609375, 0.376953125, 0.215576171875, -0.1619873046875, 0.183837890625, -0.056396484375, 0.3349609375, 0.340087890625, 0.46337890625, -0.233642578125, -0.2294921875, 0.150146484375, -0.2276611328125, 0.01953125, -0.31103515625, 0.3515625, 0.45068359375, -0.30322265625, 0.439697265625, -2.126953125, 0.52734375, 0.489501953125, 0.104248046875, -0.5595703125, -0.1619873046875, 0.1065673828125, -0.1591796875, 0.11737060546875, 0.15185546875, 0.43310546875, -0.142822265625, 0.8779296875, -0.317138671875, -0.0400390625, 0.59228515625, 0.2646484375, 0.2392578125, -0.1923828125, -0.38525390625, 0.206787109375, 0.47021484375, 0.1566162109375, 0.01922607421875, -0.282958984375, 0.332275390625, 0.1781005859375, 1.015625, 0.25830078125, 0.321533203125, -0.132568359375, 0.360107421875, 0.140869140625, 0.051513671875, 0.6611328125, -0.767578125, -0.365966796875, 0.51806640625, -0.56591796875, 0.010986328125, 0.82861328125, -0.41259765625, -0.2646484375, -0.13427734375, -0.47607421875, -0.18701171875, 0.12841796875, -0.978515625, -0.02630615234375, 0.1256103515625, -0.042724609375, 0.64208984375, -0.27392578125, 0.380859375, -0.353515625, -0.2509765625, -0.1802978515625, -0.147705078125, 0.02685546875, -0.2197265625, 0.1307373046875, 0.10040283203125, 0.08740234375, 0.233642578125, -0.198974609375, -0.1748046875, -0.389404296875, 0.01074981689453125, 0.53955078125, 1.2509765625, -0.205078125, -0.301025390625, -0.552734375, 0.5234375, 0.388671875, 0.09454345703125, 0.3037109375, -0.244140625, -1.2724609375, -0.09521484375, 0.0208740234375, -0.6064453125, 0.41015625, 0.52783203125, -0.07550048828125, 0.1591796875, -0.1234130859375, 0.09814453125, 0.21826171875, 0.0482177734375, -0.48779296875, 0.26123046875, -0.1292724609375, -0.418212890625, 0.16943359375, -0.1240234375, -0.309326171875, -0.3056640625, -0.1868896484375, -0.08935546875, -0.48291015625, 0.066650390625, 0.010009765625, -0.5224609375, -0.4912109375, -0.213623046875, -0.47314453125, -0.0203857421875, 0.207763671875, -0.035400390625, -0.32275390625, -0.2412109375, -0.154296875, -0.2783203125, 0.2314453125, 0.431640625, 0.134521484375, 0.21533203125, -0.79833984375, -0.348876953125, -0.5517578125, -0.22119140625, 0.0025634765625, 0.2005615234375, 0.1859130859375, 0.109375, 0.467529296875, -0.048858642578125, 0.1854248046875, -0.0638427734375, -0.1968994140625, 0.39599609375, 0.262451171875, 0.603515625, -0.2060546875, 0.068359375, -0.80029296875, -0.1243896484375, 0.0543212890625, 0.2376708984375, -0.2607421875, -0.084716796875, -0.541015625, 0.2783203125, -1.37109375, 0.150146484375, -0.08953857421875, 1.2646484375, -0.13134765625, 0.386962890625, 0.368896484375, 0.430419921875, -0.114013671875, 0.01702880859375, 0.28759765625, -0.09228515625, 0.337646484375, 0.62890625, -0.6396484375, -0.222900390625, -0.182861328125, -0.17822265625, -0.201416015625, 0.159423828125, 0.50341796875, 9.1552734375e-05, -0.59765625, 0.18115234375, -0.202880859375, -0.03924560546875, -0.8037109375, 0.83837890625, 0.0531005859375, -0.056884765625, 0.006591796875, -0.8056640625, 0.46826171875, 0.1485595703125, 0.54345703125, -0.019775390625, -0.6552734375, 0.625, 0.28173828125, -0.10400390625, 1.130859375, 0.0191650390625, -0.0634765625, -0.6044921875, -0.2169189453125, -0.234375, 0.0440673828125, 0.072998046875, 0.087158203125, 0.498046875, -0.154052734375, 0.148681640625, 0.0211181640625, -0.5361328125, -0.323974609375, 0.302734375, -0.38134765625, -0.2008056640625, 0.11749267578125, -0.42138671875, -0.1265869140625, -0.026824951171875, -0.10009765625, 0.87158203125, -0.0103759765625, -0.2406005859375, -0.151123046875, 0.171630859375, 0.1427001953125, -0.0333251953125, 0.31298828125, -0.17919921875, 0.2222900390625, 0.6201171875, -0.46630859375, 0.17333984375, -0.292724609375, -0.83349609375, 0.21728515625, 0.2490234375, 0.114501953125, -0.04022216796875, 0.0565185546875, 0.1513671875, 0.10693359375, -0.191650390625, 0.3779296875, 0.72509765625]}, {"frame_number": 17, "vector": [-0.172119140625, 0.2239990234375, -0.0234375, -0.1358642578125, -0.642578125, 0.341796875, -0.0115966796875, -0.4580078125, -0.11328125, -0.05780029296875, -0.115478515625, 0.060302734375, 0.23583984375, 0.255859375, 0.001953125, 0.302734375, 0.0987548828125, -0.15771484375, -0.0821533203125, -0.5517578125, -0.5537109375, -6.16015625, -0.395751953125, -0.72265625, 0.049560546875, -0.1209716796875, -0.2236328125, -0.8740234375, 0.459228515625, -0.161376953125, -0.9619140625, -0.58349609375, -0.66796875, 0.034423828125, -0.0833740234375, -0.011627197265625, 0.352783203125, -0.625, -0.0750732421875, -0.439697265625, -0.0107421875, 0.2374267578125, 0.204833984375, -0.011474609375, 0.98046875, -0.418212890625, -0.3876953125, -0.175537109375, -0.546875, -0.16845703125, 0.0003204345703125, -0.283935546875, 0.14208984375, 0.52734375, 0.187744140625, -0.02001953125, 0.42333984375, -0.281494140625, 0.30126953125, 0.0107421875, -0.073486328125, 0.173828125, 0.259033203125, 0.193359375, -0.2171630859375, -0.28564453125, -0.50048828125, -0.01708984375, 0.273681640625, 0.07891845703125, 0.9501953125, -0.0262451171875, 0.018798828125, -0.05810546875, -0.377197265625, 0.8359375, -0.279296875, -0.40625, 0.241455078125, -0.6640625, 1.193359375, 0.287841796875, 0.6083984375, 0.5791015625, -0.26953125, 0.05029296875, 0.2254638671875, -0.05377197265625, 0.15673828125, -0.623046875, -0.1219482421875, -0.12457275390625, 0.0389404296875, -0.023193359375, 0.307373046875, 0.3369140625, 0.10333251953125, 0.54638671875, 0.06622314453125, -0.62841796875, 0.64453125, 0.365966796875, -0.35009765625, -0.6064453125, 0.01666259765625, -0.0838623046875, 0.10491943359375, -0.09600830078125, -0.1376953125, -0.01171875, -0.0750732421875, 0.152587890625, 0.042938232421875, -0.385009765625, 0.115234375, -0.583984375, -0.13134765625, -0.1488037109375, 0.341064453125, 0.5556640625, 0.058837890625, -0.259765625, -0.0333251953125, -0.59326171875, 0.0191650390625, 0.28466796875, -0.1409912109375, 0.423828125, 0.130615234375, 0.250244140625, -0.04833984375, -0.429931640625, -0.0706787109375, 0.04534912109375, 0.509765625, -0.099609375, 0.12103271484375, 0.79833984375, 0.556640625, 0.07373046875, -0.03564453125, 0.27392578125, 0.172119140625, -0.120849609375, -0.32763671875, 0.1478271484375, 0.013702392578125, -0.38623046875, 0.2044677734375, 1.099609375, -0.1204833984375, -1.6328125, 0.1414794921875, -0.4775390625, 0.3876953125, -0.258544921875, -0.1533203125, 0.94384765625, 0.3017578125, -0.4462890625, -0.262939453125, 0.169189453125, 0.39990234375, 0.337158203125, -0.29345703125, -0.26171875, 0.0311279296875, 0.6552734375, 0.034423828125, 0.0072021484375, 0.2425537109375, 0.43994140625, -0.143310546875, 0.6015625, -0.263916015625, 0.96923828125, -1.06640625, 0.1092529296875, 0.0081787109375, -0.080322265625, 0.073486328125, 0.154052734375, -0.19970703125, -0.0200653076171875, 0.0020751953125, 0.01434326171875, -0.0185546875, 0.18115234375, -0.069580078125, 0.11669921875, 0.155029296875, 0.142578125, 0.476806640625, -0.419921875, 0.5419921875, -0.28515625, 0.442138671875, -0.274658203125, -0.2353515625, 1.130859375, 0.517578125, -0.02490234375, -0.266357421875, -0.014404296875, -0.00872802734375, 0.1009521484375, 0.71484375, 0.048370361328125, -0.65625, -0.165771484375, 0.1995849609375, -0.081298828125, 0.204833984375, -0.87158203125, -0.50146484375, -0.525390625, 0.53515625, 1.0234375, -0.0021514892578125, -0.50390625, -0.48681640625, -0.52099609375, -0.0926513671875, -0.6162109375, -0.1650390625, 0.23486328125, -0.07415771484375, -0.2333984375, 0.6865234375, -0.0042724609375, -0.109619140625, 0.237060546875, -0.01513671875, 0.3095703125, -0.265625, -0.63330078125, 0.6962890625, -0.1451416015625, -0.96435546875, -0.056640625, 0.72802734375, 0.311767578125, -0.09844970703125, 0.22998046875, -0.26708984375, 0.98779296875, -0.23876953125, -0.0015869140625, -0.264404296875, -0.19580078125, 0.0379638671875, -0.02801513671875, 0.317626953125, -0.0771484375, -0.662109375, -0.06494140625, -0.10498046875, -0.063232421875, 0.408447265625, 0.479736328125, -0.023193359375, -0.19775390625, -0.0303955078125, -0.0421142578125, -0.317626953125, 0.05194091796875, -0.7802734375, -0.32373046875, 0.5234375, -0.41748046875, 0.05804443359375, 0.484130859375, 0.36328125, 0.24462890625, -0.2646484375, -0.712890625, -0.1329345703125, 0.1328125, -0.323974609375, -0.416015625, -0.297607421875, 0.658203125, -0.4677734375, 0.166015625, 0.45068359375, 0.310791015625, -0.153564453125, 0.537109375, -0.5556640625, 1.126953125, -0.004852294921875, 0.055023193359375, 0.08154296875, -0.419677734375, 0.0494384765625, 0.1156005859375, 0.182861328125, 0.3994140625, 0.5400390625, -0.240478515625, 0.0193939208984375, 0.787109375, -0.01335906982421875, -0.032958984375, -0.374755859375, 0.404296875, 0.4169921875, -0.6748046875, 0.1307373046875, -0.117919921875, 0.28759765625, -0.44873046875, 0.034912109375, -0.271484375, -0.28369140625, 0.38671875, -0.0655517578125, 0.06396484375, -0.2078857421875, 0.154052734375, -0.9228515625, 0.176513671875, 0.0323486328125, -0.400390625, -0.8828125, 0.52880859375, -0.439697265625, 0.2724609375, -0.55615234375, -0.080810546875, -0.457275390625, -0.0477294921875, -0.1728515625, -0.48291015625, 0.1356201171875, 0.29296875, 0.1048583984375, -0.6787109375, 0.02886962890625, -0.041259765625, -0.693359375, 0.137939453125, -0.291748046875, -0.6796875, -0.58935546875, 0.166015625, -0.419921875, 0.33984375, 0.219482421875, -0.0311279296875, -0.3896484375, -0.41064453125, 0.2822265625, 0.15771484375, 0.4091796875, 0.306640625, 0.23876953125, -0.107177734375, -0.265625, -0.2476806640625, -0.21923828125, -0.0032958984375, 0.46923828125, 0.21044921875, -0.607421875, -0.5107421875, -0.021484375, 0.24853515625, -0.01055908203125, -0.10748291015625, -0.02972412109375, 0.2493896484375, -0.13916015625, 0.371337890625, 0.354736328125, -0.14111328125, 0.3583984375, -0.34814453125, 0.0869140625, -0.61474609375, -0.046875, -0.269287109375, 0.1240234375, 0.7060546875, 0.09564208984375, -0.055419921875, -0.060760498046875, -0.064697265625, -0.2462158203125, 0.5361328125, -0.5546875, -0.19580078125, 0.2421875, 0.275146484375, 1.029296875, 0.6279296875, -0.138671875, -0.681640625, -0.0601806640625, -0.25244140625, 0.217041015625, 0.147705078125, -0.1171875, -0.3984375, -0.165771484375, 0.08447265625, -0.359375, 0.033935546875, 0.345458984375, -0.0723876953125, 0.33154296875, -0.884765625, -0.7080078125, -0.25634765625, -0.416748046875, 0.5693359375, 0.79248046875, -0.32861328125, 1.107421875, -0.2374267578125, 0.2191162109375, -0.051361083984375, -0.265625, -0.093994140625, -0.00390625, -0.53759765625, -1.3505859375, -0.0408935546875, 0.1748046875, 0.181884765625, 0.04937744140625, -6.1796875, -0.364013671875, -0.041259765625, -0.67626953125, -0.29443359375, -0.0777587890625, -1.251953125, -0.66552734375, -0.25634765625, -0.04107666015625, 0.5419921875, 0.568359375, -0.296875, 0.0560302734375, 0.27001953125, 0.219970703125, -0.0072021484375, -0.154052734375, 0.2193603515625, -0.039794921875, 0.1942138671875, -0.11767578125, 0.155517578125, 0.443603515625, -0.280029296875, -0.168701171875, 0.2900390625, -0.640625, -0.64404296875, -0.151123046875, 0.130615234375, -0.318603515625, -0.2998046875, -0.470703125, -0.0787353515625, 0.59130859375, 0.8037109375, -0.337890625, -0.4013671875, 0.2197265625, -1.02734375, 0.29931640625, -0.031982421875, -0.0836181640625, 0.212158203125, 0.405029296875, -0.09405517578125, -0.75927734375, -0.77734375, 0.0931396484375, 0.02294921875, -1.064453125, -0.092041015625, 0.33935546875, -0.1060791015625, 0.25830078125, -0.1185302734375, 0.012054443359375, 0.172607421875, 0.21484375, 0.4833984375, 0.28662109375, 0.3818359375, -0.389892578125, 0.052001953125, 0.08453369140625, 0.0028076171875, 0.110595703125, 0.0155029296875, -0.120361328125, 0.014801025390625, 0.6875, -0.10687255859375, -0.37744140625, -0.099609375, -0.1083984375, 0.1514892578125, -0.0438232421875, 0.1025390625, 0.199462890625, 0.0631103515625, 0.07843017578125, 0.3115234375, -0.24853515625, -0.1917724609375, -0.060546875, -0.400634765625, -0.61865234375, -0.09716796875, 0.755859375, 0.179443359375, -0.2052001953125, 0.11962890625, 0.2216796875, 1.251953125, 0.166259765625, 0.18896484375, 0.09869384765625, 0.143310546875, -0.447265625, -0.453369140625, 0.020751953125, -0.0283203125, 0.342529296875, 0.2086181640625, 0.063232421875, -0.03570556640625, 0.3994140625, 0.091064453125, 0.7119140625, 0.1087646484375, 0.398681640625, 0.52783203125, -0.13671875, 0.5703125, -1.5322265625, -0.3837890625, -0.288818359375, 0.1295166015625, -0.056640625, -0.3525390625, 0.74169921875, 0.310302734375, -0.095703125, 0.081787109375, 0.3798828125, -0.1021728515625, -0.23291015625, -0.98388671875, -0.46826171875, -0.34375, -0.40234375, -0.91748046875, 0.3935546875, 0.2467041015625, 0.560546875, -0.45263671875, -0.121826171875, 0.1151123046875, -0.091552734375, 0.36962890625, -0.244140625, 0.07391357421875, 0.22119140625, 0.4755859375, -0.4384765625, 0.20556640625, -0.302490234375, -0.55029296875, -0.30078125, 0.0927734375, -0.189208984375, -0.1212158203125, 0.01446533203125, -0.065673828125, 0.03369140625, 0.47119140625, -0.232666015625, 0.2420654296875, 0.436767578125, 0.0762939453125, -0.115478515625, 0.07916259765625, 0.119140625, 0.1373291015625, -0.13134765625, -0.239990234375, -1.560546875, -0.478515625, 0.27978515625, -0.04217529296875, -0.36181640625, -0.3759765625, -0.20263671875, -0.09033203125, -0.05780029296875, -0.224365234375, -0.249755859375, -0.86865234375, 0.1907958984375, 0.35009765625, -0.2330322265625, -0.198486328125, -0.189453125, -0.552734375, 0.269287109375, 0.283447265625, -0.00067138671875, -0.244873046875, 0.072998046875, -0.078125, 0.050689697265625, 0.06024169921875, -0.328369140625, 0.00225830078125, 0.044281005859375, 0.34326171875, 0.216796875, -0.052001953125, 0.1748046875, -0.0428466796875, -0.375, 0.6572265625, 0.02667236328125, -0.140380859375, -0.35986328125, 1.330078125, 0.10791015625, -0.317626953125, 0.02734375, -0.37158203125, 0.357421875, -0.1954345703125, -0.175537109375, 0.2919921875, 0.1646728515625, 0.0069580078125, 0.260498046875, 0.1002197265625, 0.2445068359375, -0.29541015625, 0.360107421875, 0.270263671875, -0.673828125, 0.1533203125, 0.271240234375, -0.0972900390625, -0.244140625, -0.165771484375, -0.31982421875, -0.80908203125, -0.01739501953125, -0.12103271484375, -0.1165771484375, -0.311279296875, 0.123046875, -0.089599609375, 0.461669921875, -0.397216796875, -1.4453125, -0.1336669921875, -0.060302734375, 0.1796875, 0.1866455078125, -0.0758056640625, 0.27099609375, 0.018798828125, -0.078125, 0.330078125, -0.019287109375, 0.232421875, 0.47265625, 0.072509765625, 0.099853515625, -0.197998046875, 0.3115234375, 0.2164306640625, -0.37744140625, -0.68310546875, -0.1680908203125, -0.378662109375, 0.033447265625, 0.29248046875, -0.233642578125, 0.78515625, -0.0765380859375, 0.0740966796875, -0.8818359375, 0.04144287109375, -0.0135498046875, 0.163818359375, -0.5771484375, 0.259521484375, -0.62109375, 0.42333984375, -0.0521240234375, 0.27685546875, -0.103271484375, -0.48779296875, -0.6826171875, -0.018829345703125, -0.5146484375, 0.478515625, -0.227783203125, 0.99951171875, -0.27099609375, 0.137939453125, -0.1201171875, -0.07208251953125, -0.4775390625, 0.08544921875, 1.4755859375, 0.030181884765625, 0.29931640625, -0.0498046875, 0.42724609375, -0.9599609375, -0.1572265625, 0.5498046875, 0.0914306640625, -0.2423095703125, 0.587890625, 0.1268310546875, 0.115966796875, 0.19775390625, -0.0205841064453125, 0.16552734375, -0.0908203125, 0.4775390625, -0.57666015625, 0.101806640625, -0.60107421875, 0.1090087890625, 0.075439453125, -0.4091796875, -0.024169921875, -0.083251953125, -0.1544189453125, 0.130126953125, -0.8271484375, -0.26611328125, 0.324462890625, -0.0858154296875, 0.225341796875, 0.033935546875, -0.114990234375, -0.7802734375, 0.0120849609375, 0.26806640625, 0.0045166015625, 0.18212890625, -0.28076171875, -0.5615234375, 0.4951171875, -0.430908203125, 0.088623046875, -0.57275390625, 0.04644775390625, -0.39306640625, 0.093505859375, 0.09521484375, 0.03955078125, 0.043487548828125, 0.260498046875, -2.345703125, 0.2890625, -0.14208984375, 0.235107421875, 0.63623046875, 0.0032958984375, -0.34423828125, -0.204833984375, 0.11279296875, -0.1202392578125, 0.23095703125, 0.06689453125, 0.137939453125, 0.499267578125, 0.31884765625, 0.3662109375, 0.061767578125, 0.26513671875, 0.6298828125, -0.27001953125, 0.390380859375, 0.04248046875, 0.0535888671875, -0.344970703125, 0.082763671875, 0.2392578125, 0.32470703125, -0.68505859375, 0.552734375, 0.1009521484375, -0.082275390625, 0.0382080078125, -0.4501953125, -0.0338134765625, 0.0419921875, 0.37060546875, -0.320068359375, 0.0118408203125, -0.18896484375, -0.153076171875, 0.4267578125, -0.029052734375, -0.33837890625, -0.703125, -0.3662109375, 0.3037109375, 0.0306396484375, -0.1964111328125, 0.349609375, -0.10687255859375, 0.281494140625, -0.44970703125, -0.239501953125, 0.1728515625, -0.36669921875, -0.40966796875, -0.615234375, -0.55908203125, 0.007568359375, 0.78857421875, 0.71142578125, -0.29833984375, 0.03564453125, 0.6484375, -0.6279296875, 0.20751953125, -0.6728515625, -0.6875, -0.48291015625, 0.14990234375, 0.08941650390625, -0.33447265625, -0.2410888671875, 0.113525390625, -0.17431640625, 0.1943359375, 0.43994140625, 0.1219482421875, 0.1749267578125, 0.051727294921875, -0.5185546875, 0.103515625, -0.092529296875, -0.06494140625, 0.73046875, 0.126220703125, -0.3017578125, 0.4345703125, -0.6064453125, 0.45703125, -0.239990234375, 0.55078125, -0.174560546875, 0.062744140625, 0.1649169921875, -0.97802734375, -0.12890625, 0.075439453125, 0.1397705078125, -0.319580078125, -0.36962890625, 0.0074462890625, 1.080078125, 0.409912109375, -0.60107421875, -0.22607421875, 0.6845703125, -0.181640625, -0.150146484375, -0.053558349609375, -0.453125, 0.253173828125, -0.1640625, -0.50927734375, 0.43115234375, 0.44921875, -0.28857421875, -0.572265625, 0.057769775390625, 0.07275390625, -0.080078125, 0.151123046875, -0.305419921875, -0.138427734375, 0.256103515625, 0.259033203125, 0.4150390625, 0.357421875, 0.23486328125, -0.089599609375, 0.236083984375, 0.0010986328125, 0.328125, 0.346923828125, 0.49462890625, -0.2421875, -0.136962890625, 0.156982421875, -0.28076171875, -0.09375, -0.295654296875, 0.39697265625, 0.453857421875, -0.2491455078125, 0.5048828125, -2.123046875, 0.45849609375, 0.4814453125, 0.1669921875, -0.470947265625, -0.169677734375, 0.1339111328125, -0.193603515625, 0.04815673828125, 0.14306640625, 0.53173828125, -0.131591796875, 0.80078125, -0.301025390625, 0.0443115234375, 0.63427734375, 0.26171875, 0.2626953125, -0.25341796875, -0.36669921875, 0.1629638671875, 0.4052734375, 0.1427001953125, -0.0074462890625, -0.31396484375, 0.3115234375, 0.114990234375, 0.98681640625, 0.2724609375, 0.30908203125, -0.303466796875, 0.3974609375, 0.163330078125, 0.00994873046875, 0.7861328125, -0.7548828125, -0.347900390625, 0.46630859375, -0.57861328125, -0.05224609375, 0.775390625, -0.46533203125, -0.2442626953125, -0.1279296875, -0.43017578125, -0.08447265625, 0.1099853515625, -1.0126953125, -0.027099609375, 0.117431640625, 0.00244140625, 0.658203125, -0.2919921875, 0.423828125, -0.289306640625, -0.26611328125, -0.186279296875, -0.105224609375, 0.045166015625, -0.18310546875, 0.1744384765625, 0.068115234375, 0.06756591796875, 0.29296875, -0.0927734375, -0.12548828125, -0.42822265625, 0.01520538330078125, 0.59326171875, 1.3447265625, -0.1505126953125, -0.2001953125, -0.5400390625, 0.53125, 0.33984375, 0.0855712890625, 0.26904296875, -0.281982421875, -1.333984375, -0.0997314453125, 0.080078125, -0.6357421875, 0.403564453125, 0.580078125, -0.056884765625, 0.213134765625, -0.12237548828125, 0.0643310546875, 0.173828125, 0.029541015625, -0.599609375, 0.23583984375, -0.12158203125, -0.5078125, 0.193115234375, -0.0899658203125, -0.3193359375, -0.1929931640625, -0.1431884765625, -0.14697265625, -0.456298828125, 0.109619140625, -0.000244140625, -0.4345703125, -0.42578125, -0.201416015625, -0.4677734375, -0.11761474609375, 0.3173828125, -0.1522216796875, -0.30810546875, -0.226806640625, -0.193603515625, -0.23486328125, 0.166015625, 0.431640625, 0.0426025390625, 0.22900390625, -0.7431640625, -0.31396484375, -0.5185546875, -0.2021484375, 0.015380859375, 0.138427734375, 0.199462890625, 0.06500244140625, 0.433349609375, -0.0307769775390625, 0.112060546875, -0.02423095703125, -0.25732421875, 0.46484375, 0.159912109375, 0.591796875, -0.147705078125, 0.046875, -0.72265625, -0.085205078125, 0.018310546875, 0.177490234375, -0.2366943359375, -0.08905029296875, -0.5244140625, 0.316650390625, -1.3837890625, 0.108154296875, -0.07952880859375, 1.2314453125, -0.14599609375, 0.430419921875, 0.3623046875, 0.3798828125, -0.2166748046875, -0.015960693359375, 0.237060546875, -0.12127685546875, 0.38916015625, 0.6591796875, -0.52685546875, -0.177490234375, -0.18359375, -0.11322021484375, -0.10101318359375, 0.17724609375, 0.447265625, 0.02374267578125, -0.546875, 0.2587890625, -0.195556640625, -0.0098876953125, -0.705078125, 0.892578125, 0.0216064453125, -0.037841796875, -0.0321044921875, -0.71875, 0.438720703125, 0.28173828125, 0.5361328125, -0.1387939453125, -0.74853515625, 0.63916015625, 0.28466796875, -0.11151123046875, 1.0927734375, -0.0064697265625, -0.127197265625, -0.6875, -0.141357421875, -0.18896484375, -0.0267333984375, 0.06005859375, 0.1458740234375, 0.49072265625, -0.17822265625, 0.06341552734375, 0.074462890625, -0.5263671875, -0.25048828125, 0.30126953125, -0.4189453125, -0.25537109375, 0.12109375, -0.411865234375, -0.07366943359375, -0.0377197265625, 0.0189208984375, 0.8759765625, 0.06103515625, -0.246337890625, -0.138671875, 0.1566162109375, 0.1544189453125, -0.00042724609375, 0.329345703125, -0.1630859375, 0.2034912109375, 0.5380859375, -0.4267578125, 0.205322265625, -0.307373046875, -0.8232421875, 0.1627197265625, 0.275146484375, 0.1591796875, 0.01519775390625, 0.1014404296875, 0.071044921875, 0.0830078125, -0.202880859375, 0.41650390625, 0.68212890625]}, {"frame_number": 18, "vector": [-0.1331787109375, 0.2734375, 0.001708984375, -0.138671875, -0.6484375, 0.31787109375, -0.042938232421875, -0.447265625, -0.0966796875, -0.103271484375, -0.139404296875, 0.044189453125, 0.30419921875, 0.181640625, -0.03173828125, 0.2705078125, 0.1033935546875, -0.164306640625, -0.0924072265625, -0.47216796875, -0.5126953125, -6.203125, -0.33447265625, -0.69677734375, 0.021636962890625, -0.10968017578125, -0.19384765625, -0.8681640625, 0.401611328125, -0.135009765625, -0.9296875, -0.61376953125, -0.7041015625, 0.05975341796875, -0.10296630859375, 0.03778076171875, 0.33935546875, -0.677734375, -0.0401611328125, -0.435791015625, -0.01849365234375, 0.209228515625, 0.203857421875, 0.01708984375, 0.994140625, -0.39697265625, -0.4072265625, -0.1925048828125, -0.5400390625, -0.1533203125, 0.0015716552734375, -0.2861328125, 0.07391357421875, 0.5537109375, 0.201416015625, 0.020538330078125, 0.42041015625, -0.27587890625, 0.32470703125, 0.019287109375, -0.11083984375, 0.13720703125, 0.2425537109375, 0.18408203125, -0.2403564453125, -0.34814453125, -0.51025390625, -0.007568359375, 0.2421875, 0.1480712890625, 0.94921875, -0.00994873046875, 0.01513671875, -0.066650390625, -0.356689453125, 0.78271484375, -0.33251953125, -0.380859375, 0.185302734375, -0.708984375, 1.19140625, 0.302734375, 0.57861328125, 0.5595703125, -0.2666015625, 0.08740234375, 0.1981201171875, -0.0120849609375, 0.189453125, -0.63232421875, -0.14013671875, -0.099609375, 0.0224609375, -0.095458984375, 0.3408203125, 0.34814453125, 0.08544921875, 0.52783203125, 0.069091796875, -0.67578125, 0.63671875, 0.3388671875, -0.33740234375, -0.59326171875, 0.016693115234375, -0.11273193359375, 0.03448486328125, -0.11663818359375, -0.114013671875, 0.012939453125, -0.0687255859375, 0.160888671875, 0.06683349609375, -0.400634765625, 0.165771484375, -0.583984375, -0.138671875, -0.1546630859375, 0.320556640625, 0.56005859375, 0.05419921875, -0.28076171875, 0.002197265625, -0.58544921875, -0.0345458984375, 0.248291015625, -0.1337890625, 0.390625, 0.1876220703125, 0.241455078125, -0.01513671875, -0.461669921875, -0.0570068359375, -0.002410888671875, 0.52978515625, -0.08544921875, 0.11865234375, 0.8037109375, 0.5478515625, 0.06689453125, -0.015869140625, 0.31005859375, 0.1746826171875, -0.0946044921875, -0.320556640625, 0.1370849609375, -0.02825927734375, -0.369384765625, 0.2176513671875, 1.111328125, -0.134033203125, -1.61328125, 0.1837158203125, -0.455078125, 0.41064453125, -0.28857421875, -0.1416015625, 1.00390625, 0.379150390625, -0.4052734375, -0.291015625, 0.1865234375, 0.37548828125, 0.37939453125, -0.27490234375, -0.28955078125, 0.0093994140625, 0.705078125, 0.0540771484375, 0.0421142578125, 0.249267578125, 0.46923828125, -0.195068359375, 0.51611328125, -0.348876953125, 0.896484375, -1.072265625, 0.036865234375, 0.0247802734375, -0.059326171875, 0.10205078125, 0.15771484375, -0.1905517578125, 0.03594970703125, -0.0015869140625, 0.0201416015625, 0.008056640625, 0.19482421875, -0.062255859375, 0.10430908203125, 0.19140625, 0.153564453125, 0.481689453125, -0.37744140625, 0.607421875, -0.26904296875, 0.385986328125, -0.2861328125, -0.3076171875, 1.130859375, 0.521484375, -0.04052734375, -0.232666015625, 0.0205078125, -0.0965576171875, 0.1085205078125, 0.71875, 0.01263427734375, -0.68701171875, -0.13330078125, 0.156005859375, -0.1044921875, 0.25732421875, -0.8779296875, -0.51318359375, -0.5302734375, 0.53125, 0.98388671875, 0.0116119384765625, -0.471923828125, -0.479248046875, -0.525390625, -0.15625, -0.59326171875, -0.157470703125, 0.218994140625, -0.08477783203125, -0.23193359375, 0.6806640625, -0.0078125, -0.032623291015625, 0.27685546875, 0.0255126953125, 0.332763671875, -0.29931640625, -0.6328125, 0.69580078125, -0.0894775390625, -0.9951171875, -0.04345703125, 0.70068359375, 0.286865234375, -0.06146240234375, 0.234619140625, -0.1871337890625, 1.01953125, -0.218994140625, -0.032470703125, -0.27880859375, -0.14453125, 0.0330810546875, -0.041015625, 0.348388671875, -0.05322265625, -0.6630859375, -0.0478515625, -0.139892578125, -0.0517578125, 0.43310546875, 0.48388671875, -0.01715087890625, -0.12042236328125, -0.080322265625, -0.040771484375, -0.29638671875, 0.02178955078125, -0.81689453125, -0.30078125, 0.5322265625, -0.4306640625, 0.0198974609375, 0.4580078125, 0.38134765625, 0.269775390625, -0.2008056640625, -0.6728515625, -0.07708740234375, 0.12091064453125, -0.33642578125, -0.43701171875, -0.26220703125, 0.65478515625, -0.400390625, 0.1796875, 0.470947265625, 0.300048828125, -0.170166015625, 0.5849609375, -0.58203125, 1.134765625, -0.04718017578125, 0.059051513671875, 0.031982421875, -0.38818359375, 0.0640869140625, 0.1561279296875, 0.1435546875, 0.4208984375, 0.564453125, -0.2451171875, 0.00281524658203125, 0.78857421875, -0.034912109375, -0.04052734375, -0.406494140625, 0.424560546875, 0.39404296875, -0.62890625, 0.126708984375, -0.076171875, 0.3037109375, -0.388671875, 0.0263671875, -0.3251953125, -0.323486328125, 0.421875, -0.0001220703125, 0.0419921875, -0.25537109375, 0.15869140625, -0.9072265625, 0.1796875, 0.07427978515625, -0.453125, -0.94677734375, 0.57568359375, -0.40234375, 0.25537109375, -0.54296875, -0.09149169921875, -0.475341796875, -0.0611572265625, -0.0997314453125, -0.47412109375, 0.22412109375, 0.2685546875, 0.1263427734375, -0.67431640625, 0.02294921875, -0.067626953125, -0.66845703125, 0.0955810546875, -0.274658203125, -0.65087890625, -0.53759765625, 0.194580078125, -0.4091796875, 0.37255859375, 0.198486328125, -0.0240478515625, -0.40771484375, -0.323486328125, 0.2890625, 0.13623046875, 0.400634765625, 0.34619140625, 0.25244140625, -0.143798828125, -0.25732421875, -0.246826171875, -0.193359375, -0.0206298828125, 0.509765625, 0.225341796875, -0.638671875, -0.521484375, -0.0091552734375, 0.2259521484375, 0.06024169921875, -0.09698486328125, -0.0517578125, 0.23681640625, -0.097412109375, 0.366943359375, 0.3203125, -0.1431884765625, 0.362548828125, -0.32861328125, 0.03955078125, -0.5849609375, -0.030517578125, -0.28125, 0.0902099609375, 0.69287109375, 0.159423828125, -0.0472412109375, -0.0318603515625, -0.076171875, -0.2205810546875, 0.54150390625, -0.580078125, -0.143310546875, 0.237548828125, 0.315673828125, 1.03515625, 0.630859375, -0.1181640625, -0.73828125, -0.0335693359375, -0.263671875, 0.1865234375, 0.171142578125, -0.111083984375, -0.37841796875, -0.196044921875, 0.04541015625, -0.406982421875, 0.04296875, 0.35693359375, -0.15673828125, 0.36474609375, -0.8828125, -0.6787109375, -0.272216796875, -0.355224609375, 0.5908203125, 0.76953125, -0.31982421875, 1.15234375, -0.2802734375, 0.236328125, -0.05389404296875, -0.25244140625, -0.11572265625, -0.0107421875, -0.466552734375, -1.333984375, -0.0428466796875, 0.13427734375, 0.1788330078125, 0.10302734375, -6.203125, -0.41162109375, -0.046630859375, -0.724609375, -0.28369140625, -0.1431884765625, -1.18359375, -0.67919921875, -0.24560546875, -0.007080078125, 0.58447265625, 0.5908203125, -0.252685546875, 0.02105712890625, 0.26025390625, 0.182373046875, 0.0247802734375, -0.19970703125, 0.179443359375, -0.019775390625, 0.213134765625, -0.06201171875, 0.18408203125, 0.425537109375, -0.297607421875, -0.1575927734375, 0.327880859375, -0.650390625, -0.59375, -0.136962890625, 0.0673828125, -0.31201171875, -0.31201171875, -0.45703125, -0.111572265625, 0.541015625, 0.8046875, -0.282958984375, -0.48974609375, 0.251953125, -1.0625, 0.30712890625, -0.0426025390625, -0.06982421875, 0.2088623046875, 0.30712890625, -0.12188720703125, -0.654296875, -0.83349609375, 0.12457275390625, 0.03497314453125, -1.064453125, -0.03466796875, 0.310791015625, -0.08447265625, 0.28125, -0.0941162109375, 0.038818359375, 0.140869140625, 0.25146484375, 0.498291015625, 0.302978515625, 0.39306640625, -0.44970703125, 0.02099609375, 0.04193115234375, 0.0009765625, 0.08544921875, -0.00238037109375, -0.0821533203125, 0.0124359130859375, 0.6494140625, -0.134765625, -0.4228515625, -0.104248046875, -0.1259765625, 0.1756591796875, -0.052734375, 0.074951171875, 0.18603515625, 0.0469970703125, 0.1287841796875, 0.2998046875, -0.21630859375, -0.22265625, -0.109375, -0.371337890625, -0.60205078125, -0.0762939453125, 0.72216796875, 0.116943359375, -0.267333984375, 0.08074951171875, 0.249755859375, 1.236328125, 0.181396484375, 0.23291015625, 0.11859130859375, 0.099853515625, -0.38330078125, -0.454833984375, 0.026123046875, -0.0185546875, 0.371826171875, 0.279541015625, 0.058837890625, -0.06805419921875, 0.43310546875, 0.063720703125, 0.64892578125, 0.123779296875, 0.35888671875, 0.529296875, -0.1763916015625, 0.5498046875, -1.5654296875, -0.356689453125, -0.26611328125, 0.1341552734375, -0.055419921875, -0.334228515625, 0.78173828125, 0.30419921875, -0.0650634765625, 0.0439453125, 0.38232421875, -0.07861328125, -0.23681640625, -1.0029296875, -0.44189453125, -0.318115234375, -0.40185546875, -0.8876953125, 0.411865234375, 0.219970703125, 0.5439453125, -0.404296875, -0.103515625, 0.1478271484375, -0.1015625, 0.378662109375, -0.255859375, 0.1168212890625, 0.174560546875, 0.469970703125, -0.41650390625, 0.206787109375, -0.298828125, -0.50244140625, -0.310546875, 0.088134765625, -0.156005859375, -0.1202392578125, 0.01458740234375, -0.05987548828125, 0.0244140625, 0.466552734375, -0.2646484375, 0.2144775390625, 0.442138671875, 0.118896484375, -0.183349609375, 0.031341552734375, 0.0611572265625, 0.1494140625, -0.10009765625, -0.305908203125, -1.5234375, -0.47900390625, 0.2325439453125, -0.03515625, -0.352294921875, -0.387451171875, -0.2401123046875, -0.069580078125, -0.084228515625, -0.265380859375, -0.245361328125, -0.8408203125, 0.2149658203125, 0.3525390625, -0.2254638671875, -0.17529296875, -0.17041015625, -0.5703125, 0.2445068359375, 0.29150390625, -0.030517578125, -0.21875, 0.02880859375, -0.099609375, 0.03857421875, 0.0701904296875, -0.30224609375, -0.014556884765625, 0.053619384765625, 0.372802734375, 0.2587890625, -0.0596923828125, 0.183349609375, -0.0234527587890625, -0.3994140625, 0.70556640625, 0.043212890625, -0.12890625, -0.355224609375, 1.328125, 0.07666015625, -0.30615234375, 0.0302734375, -0.397216796875, 0.36181640625, -0.2342529296875, -0.14453125, 0.30908203125, 0.11572265625, -0.000213623046875, 0.2451171875, 0.06201171875, 0.231201171875, -0.307373046875, 0.343994140625, 0.260986328125, -0.64404296875, 0.180908203125, 0.271484375, -0.08221435546875, -0.22802734375, -0.2310791015625, -0.31298828125, -0.8203125, 0.0003662109375, -0.0877685546875, -0.06317138671875, -0.347412109375, 0.096923828125, -0.099365234375, 0.423583984375, -0.401123046875, -1.416015625, -0.0830078125, -0.11376953125, 0.198486328125, 0.2091064453125, -0.0634765625, 0.314697265625, 0.051025390625, -0.071044921875, 0.343994140625, -0.01800537109375, 0.250244140625, 0.47607421875, 0.09912109375, 0.126953125, -0.183349609375, 0.322998046875, 0.1546630859375, -0.397216796875, -0.62841796875, -0.184814453125, -0.3583984375, -0.028564453125, 0.31494140625, -0.22265625, 0.8662109375, -0.1025390625, 0.1234130859375, -0.91064453125, 0.03204345703125, -0.04815673828125, 0.131103515625, -0.6103515625, 0.269775390625, -0.6064453125, 0.39111328125, -0.024017333984375, 0.2998046875, -0.1435546875, -0.4912109375, -0.650390625, 0.057373046875, -0.45263671875, 0.47509765625, -0.21533203125, 0.9833984375, -0.343505859375, 0.0736083984375, -0.115966796875, -0.0709228515625, -0.462890625, 0.0107421875, 1.453125, 0.0234375, 0.285400390625, -0.080322265625, 0.41259765625, -1.046875, -0.150634765625, 0.50390625, 0.05767822265625, -0.222412109375, 0.60546875, 0.1441650390625, 0.106689453125, 0.205810546875, -0.0063018798828125, 0.1904296875, -0.097412109375, 0.48388671875, -0.58984375, 0.0703125, -0.57275390625, 0.0386962890625, 0.1112060546875, -0.49365234375, -0.0242919921875, -0.0716552734375, -0.10516357421875, 0.086181640625, -0.8203125, -0.190185546875, 0.30810546875, -0.110107421875, 0.2406005859375, 0.041259765625, -0.0958251953125, -0.828125, 0.0340576171875, 0.295166015625, 0.00885009765625, 0.13818359375, -0.258544921875, -0.55322265625, 0.45703125, -0.4609375, 0.0599365234375, -0.58984375, 0.0718994140625, -0.38330078125, 0.05767822265625, 0.03759765625, 0.07318115234375, 0.025848388671875, 0.254150390625, -2.3046875, 0.265625, -0.170166015625, 0.185546875, 0.68017578125, -0.00634765625, -0.34619140625, -0.18701171875, 0.21533203125, -0.1424560546875, 0.195068359375, 0.05615234375, 0.0594482421875, 0.4931640625, 0.35888671875, 0.3447265625, 0.007568359375, 0.250244140625, 0.615234375, -0.26806640625, 0.332275390625, 0.02398681640625, 0.06182861328125, -0.3466796875, 0.114990234375, 0.2362060546875, 0.34033203125, -0.705078125, 0.5849609375, 0.0870361328125, -0.1494140625, 0.0484619140625, -0.47705078125, -0.0272216796875, 0.051727294921875, 0.33154296875, -0.2568359375, 0.0040283203125, -0.18603515625, -0.171875, 0.412353515625, -0.03570556640625, -0.3271484375, -0.6728515625, -0.349853515625, 0.25390625, 0.0467529296875, -0.1859130859375, 0.279052734375, -0.1112060546875, 0.289306640625, -0.399169921875, -0.230224609375, 0.161865234375, -0.404541015625, -0.428466796875, -0.63037109375, -0.52197265625, 0.007843017578125, 0.7470703125, 0.6826171875, -0.2939453125, 0.0098876953125, 0.6298828125, -0.6376953125, 0.22509765625, -0.64404296875, -0.7158203125, -0.452392578125, 0.1318359375, 0.085693359375, -0.330322265625, -0.243408203125, 0.1849365234375, -0.10595703125, 0.16162109375, 0.410888671875, 0.1158447265625, 0.159423828125, 0.042633056640625, -0.49365234375, 0.09326171875, -0.0921630859375, -0.07421875, 0.72216796875, 0.129638671875, -0.31689453125, 0.422119140625, -0.611328125, 0.4619140625, -0.25341796875, 0.57421875, -0.1256103515625, 0.056396484375, 0.1751708984375, -0.9443359375, -0.1259765625, 0.0487060546875, 0.1707763671875, -0.307861328125, -0.430908203125, -0.0169677734375, 1.09375, 0.35693359375, -0.63330078125, -0.204345703125, 0.611328125, -0.140380859375, -0.1422119140625, -0.0819091796875, -0.43798828125, 0.2435302734375, -0.178955078125, -0.45947265625, 0.4365234375, 0.3837890625, -0.295654296875, -0.576171875, 0.0845947265625, 0.0828857421875, -0.068115234375, 0.12255859375, -0.26708984375, -0.09271240234375, 0.2008056640625, 0.354248046875, 0.390625, 0.368408203125, 0.243896484375, -0.10693359375, 0.1650390625, -0.060302734375, 0.345458984375, 0.339599609375, 0.54296875, -0.3154296875, -0.1395263671875, 0.1513671875, -0.2744140625, -0.09130859375, -0.2685546875, 0.4345703125, 0.415283203125, -0.21142578125, 0.501953125, -2.1171875, 0.47705078125, 0.409423828125, 0.1357421875, -0.464111328125, -0.13671875, 0.0712890625, -0.17138671875, 0.0352783203125, 0.2392578125, 0.434814453125, -0.17138671875, 0.7470703125, -0.315673828125, 0.0379638671875, 0.66552734375, 0.2393798828125, 0.26708984375, -0.2451171875, -0.38330078125, 0.2198486328125, 0.48583984375, 0.16357421875, -0.029205322265625, -0.333251953125, 0.315185546875, 0.10186767578125, 0.90625, 0.234130859375, 0.34521484375, -0.31201171875, 0.37451171875, 0.22509765625, -0.0194091796875, 0.75146484375, -0.6728515625, -0.2646484375, 0.50048828125, -0.6044921875, 0.006591796875, 0.75, -0.47900390625, -0.2705078125, -0.005126953125, -0.42626953125, -0.150146484375, 0.1324462890625, -0.9794921875, -0.072509765625, 0.10491943359375, -0.0140380859375, 0.60693359375, -0.30517578125, 0.38427734375, -0.270751953125, -0.25927734375, -0.2091064453125, -0.060546875, 0.07861328125, -0.1488037109375, 0.1380615234375, -0.014190673828125, 0.04229736328125, 0.265869140625, -0.077392578125, -0.10791015625, -0.3779296875, -0.01496124267578125, 0.56640625, 1.3515625, -0.1533203125, -0.183349609375, -0.51806640625, 0.5390625, 0.34228515625, 0.0653076171875, 0.27587890625, -0.263671875, -1.2890625, -0.095947265625, 0.0367431640625, -0.64111328125, 0.42041015625, 0.5927734375, -0.0474853515625, 0.25634765625, -0.156494140625, 0.092529296875, 0.127197265625, -0.036956787109375, -0.5234375, 0.252197265625, -0.1490478515625, -0.5068359375, 0.19384765625, -0.1044921875, -0.331787109375, -0.18505859375, -0.1064453125, -0.168212890625, -0.401611328125, 0.096923828125, -0.0010986328125, -0.3662109375, -0.4375, -0.143310546875, -0.50537109375, -0.10589599609375, 0.303955078125, -0.066162109375, -0.293212890625, -0.1943359375, -0.2099609375, -0.1932373046875, 0.15869140625, 0.49951171875, 0.00860595703125, 0.2220458984375, -0.69140625, -0.324462890625, -0.448974609375, -0.1917724609375, -0.0047607421875, 0.1756591796875, 0.2320556640625, 0.03912353515625, 0.41748046875, -0.03131103515625, 0.15966796875, -0.04345703125, -0.26708984375, 0.42822265625, 0.1463623046875, 0.5771484375, -0.1331787109375, 0.0517578125, -0.7890625, -0.0538330078125, 0.0491943359375, 0.1595458984375, -0.269775390625, -0.111572265625, -0.529296875, 0.317626953125, -1.3671875, 0.1329345703125, -0.09112548828125, 1.23046875, -0.1383056640625, 0.418212890625, 0.38427734375, 0.37255859375, -0.2374267578125, -0.057952880859375, 0.248291015625, -0.0821533203125, 0.412353515625, 0.63671875, -0.53662109375, -0.2265625, -0.1563720703125, -0.11480712890625, -0.07586669921875, 0.214599609375, 0.4287109375, 0.0562744140625, -0.5302734375, 0.20654296875, -0.21337890625, -0.059356689453125, -0.693359375, 0.9208984375, 0.013427734375, 0.025146484375, -0.009033203125, -0.724609375, 0.439453125, 0.3232421875, 0.5576171875, -0.1435546875, -0.7763671875, 0.67236328125, 0.292236328125, -0.088134765625, 1.0625, 0.0223388671875, -0.0955810546875, -0.75, -0.15771484375, -0.240478515625, -0.02392578125, 0.0576171875, 0.15234375, 0.470703125, -0.1904296875, 0.06768798828125, 0.021240234375, -0.49951171875, -0.2548828125, 0.283935546875, -0.43896484375, -0.273681640625, 0.1473388671875, -0.396484375, -0.1112060546875, -0.1278076171875, -0.0067138671875, 0.845703125, 0.03924560546875, -0.215087890625, -0.16650390625, 0.1641845703125, 0.1512451171875, -0.019775390625, 0.3046875, -0.1361083984375, 0.214599609375, 0.59326171875, -0.38134765625, 0.15576171875, -0.276611328125, -0.8408203125, 0.1383056640625, 0.23046875, 0.145751953125, 0.0029296875, 0.06103515625, 0.076904296875, 0.086181640625, -0.177001953125, 0.390625, 0.6962890625]}, {"frame_number": 19, "vector": [-0.150634765625, 0.275634765625, -0.022216796875, -0.15673828125, -0.666015625, 0.324462890625, 0.001708984375, -0.438720703125, -0.0723876953125, -0.044342041015625, -0.125732421875, 0.03466796875, 0.34130859375, 0.186767578125, -0.0062255859375, 0.27587890625, 0.130615234375, -0.151611328125, -0.064208984375, -0.451171875, -0.58447265625, -6.2265625, -0.279296875, -0.7109375, 0.03973388671875, -0.0970458984375, -0.189453125, -0.8271484375, 0.462158203125, -0.161865234375, -0.921875, -0.60595703125, -0.7021484375, 0.06201171875, -0.06793212890625, 0.0689697265625, 0.362548828125, -0.63330078125, -0.03765869140625, -0.465576171875, -0.06610107421875, 0.1812744140625, 0.1861572265625, 0.008544921875, 0.974609375, -0.37451171875, -0.420654296875, -0.1883544921875, -0.58740234375, -0.12890625, 0.0163116455078125, -0.29345703125, 0.1025390625, 0.5458984375, 0.212158203125, 0.051910400390625, 0.432373046875, -0.26513671875, 0.287109375, 0.042236328125, -0.069580078125, 0.138671875, 0.249267578125, 0.1859130859375, -0.304931640625, -0.390625, -0.487060546875, 0.041748046875, 0.24853515625, 0.1239013671875, 0.939453125, -0.0069580078125, -0.009033203125, -0.038330078125, -0.37548828125, 0.8173828125, -0.35009765625, -0.40673828125, 0.150390625, -0.7099609375, 1.177734375, 0.32421875, 0.57470703125, 0.49951171875, -0.28173828125, 0.0865478515625, 0.201416015625, -0.0028076171875, 0.1865234375, -0.63525390625, -0.13330078125, -0.0802001953125, 0.02197265625, -0.130859375, 0.325927734375, 0.36474609375, 0.078369140625, 0.47802734375, 0.0654296875, -0.671875, 0.7177734375, 0.298828125, -0.373046875, -0.64013671875, 0.03662109375, -0.090576171875, 0.02447509765625, -0.135986328125, -0.1270751953125, -0.03076171875, -0.072998046875, 0.1220703125, 0.113037109375, -0.480712890625, 0.14208984375, -0.5703125, -0.17822265625, -0.13671875, 0.289306640625, 0.5380859375, 0.0645751953125, -0.2763671875, -0.0115966796875, -0.65380859375, -0.0855712890625, 0.285888671875, -0.146240234375, 0.371337890625, 0.1802978515625, 0.273193359375, -0.0101318359375, -0.4677734375, -0.0482177734375, -0.05303955078125, 0.55859375, -0.048095703125, 0.12451171875, 0.8046875, 0.5322265625, 0.0498046875, 0.0308837890625, 0.331298828125, 0.13623046875, -0.1259765625, -0.297119140625, 0.1015625, -0.0509033203125, -0.3681640625, 0.220458984375, 1.09765625, -0.165771484375, -1.57421875, 0.1546630859375, -0.4267578125, 0.419677734375, -0.295166015625, -0.19482421875, 0.95849609375, 0.3525390625, -0.44091796875, -0.311767578125, 0.23486328125, 0.38671875, 0.422119140625, -0.27392578125, -0.298583984375, -0.0048828125, 0.6923828125, 0.052001953125, 0.07568359375, 0.169921875, 0.44970703125, -0.2054443359375, 0.560546875, -0.33154296875, 0.9794921875, -1.0673828125, 0.093017578125, -0.017578125, -0.086669921875, 0.06689453125, 0.1260986328125, -0.210205078125, 0.08966064453125, 0.0010986328125, 0.02825927734375, 0.00048828125, 0.161865234375, -0.07763671875, 0.0999755859375, 0.234375, 0.16357421875, 0.48681640625, -0.40185546875, 0.609375, -0.237548828125, 0.43017578125, -0.23291015625, -0.34130859375, 1.18359375, 0.48681640625, 0.00146484375, -0.19384765625, 0.048095703125, -0.0677490234375, 0.0887451171875, 0.6162109375, -0.0313720703125, -0.63671875, -0.11932373046875, 0.186767578125, -0.052490234375, 0.244384765625, -0.84521484375, -0.52734375, -0.5576171875, 0.483154296875, 1.01171875, -0.03204345703125, -0.419677734375, -0.458740234375, -0.54296875, -0.125, -0.61767578125, -0.145751953125, 0.211669921875, -0.113525390625, -0.208984375, 0.70849609375, 0.012939453125, -0.042266845703125, 0.289794921875, 0.0615234375, 0.35693359375, -0.342041015625, -0.5771484375, 0.7265625, -0.0350341796875, -1.021484375, -0.068603515625, 0.6826171875, 0.24658203125, -0.0260009765625, 0.17822265625, -0.1361083984375, 1.01171875, -0.1844482421875, -0.0059814453125, -0.297607421875, -0.13037109375, 0.0128173828125, -0.06378173828125, 0.353515625, -0.049560546875, -0.671875, -0.03515625, -0.13330078125, -0.042236328125, 0.43798828125, 0.45947265625, -0.042755126953125, -0.1851806640625, -0.072021484375, -0.054443359375, -0.2822265625, -0.028564453125, -0.83349609375, -0.3095703125, 0.58203125, -0.4091796875, 0.01458740234375, 0.458984375, 0.38427734375, 0.263916015625, -0.212890625, -0.70751953125, -0.15576171875, 0.126220703125, -0.35302734375, -0.392333984375, -0.269287109375, 0.6748046875, -0.40576171875, 0.1552734375, 0.420166015625, 0.312744140625, -0.158203125, 0.5927734375, -0.572265625, 1.166015625, -0.04150390625, 0.0782470703125, 0.050048828125, -0.421875, 0.08966064453125, 0.1751708984375, 0.134033203125, 0.36865234375, 0.57421875, -0.25927734375, 0.017181396484375, 0.771484375, -0.04150390625, -0.02734375, -0.32275390625, 0.402587890625, 0.37841796875, -0.65869140625, 0.090576171875, -0.060302734375, 0.3681640625, -0.368896484375, -0.012939453125, -0.329345703125, -0.37646484375, 0.3740234375, -0.01947021484375, 0.038818359375, -0.2783203125, 0.164306640625, -0.8828125, 0.139892578125, 0.13134765625, -0.4521484375, -0.93603515625, 0.5771484375, -0.4384765625, 0.27294921875, -0.6162109375, -0.0784912109375, -0.501953125, -0.097900390625, -0.1148681640625, -0.49072265625, 0.216552734375, 0.287353515625, 0.12646484375, -0.64892578125, 0.0299072265625, -0.02880859375, -0.73974609375, 0.082275390625, -0.313720703125, -0.6396484375, -0.56494140625, 0.18408203125, -0.387451171875, 0.360107421875, 0.17626953125, 0.0086669921875, -0.323974609375, -0.301513671875, 0.30126953125, 0.165283203125, 0.413818359375, 0.36865234375, 0.2412109375, -0.185302734375, -0.327392578125, -0.208984375, -0.1943359375, -0.0667724609375, 0.52783203125, 0.2578125, -0.615234375, -0.5146484375, -0.015625, 0.2313232421875, 0.067138671875, -0.09564208984375, 0.00506591796875, 0.232421875, -0.115478515625, 0.33935546875, 0.317626953125, -0.108154296875, 0.4169921875, -0.2459716796875, 0.07177734375, -0.5537109375, -0.0406494140625, -0.3349609375, 0.1082763671875, 0.71484375, 0.1895751953125, -0.0523681640625, -0.03912353515625, -0.109375, -0.2261962890625, 0.499267578125, -0.6328125, -0.118896484375, 0.287109375, 0.273193359375, 0.97900390625, 0.6376953125, -0.08251953125, -0.734375, -0.086181640625, -0.3486328125, 0.18994140625, 0.2147216796875, -0.0947265625, -0.38134765625, -0.2083740234375, 0.020751953125, -0.392822265625, 0.03558349609375, 0.37890625, -0.11669921875, 0.24853515625, -0.9013671875, -0.69580078125, -0.28173828125, -0.3486328125, 0.5244140625, 0.74365234375, -0.28125, 1.115234375, -0.29248046875, 0.2086181640625, -0.054595947265625, -0.225341796875, -0.079345703125, 0.012939453125, -0.47998046875, -1.3447265625, -0.0595703125, 0.10546875, 0.22265625, 0.06658935546875, -6.2265625, -0.38818359375, -0.072998046875, -0.79345703125, -0.2301025390625, -0.138916015625, -1.18359375, -0.65625, -0.2705078125, 0.0108642578125, 0.64892578125, 0.66015625, -0.2442626953125, -0.02154541015625, 0.21826171875, 0.17529296875, 0.0283203125, -0.1221923828125, 0.1712646484375, -0.029296875, 0.218994140625, -0.059326171875, 0.158935546875, 0.428466796875, -0.30224609375, -0.1728515625, 0.34228515625, -0.65185546875, -0.6044921875, -0.1357421875, 0.056884765625, -0.333984375, -0.31494140625, -0.46923828125, -0.0712890625, 0.546875, 0.8095703125, -0.302978515625, -0.466796875, 0.23095703125, -1.05859375, 0.33349609375, -0.0303955078125, -0.056915283203125, 0.212158203125, 0.301025390625, -0.1246337890625, -0.6943359375, -0.78662109375, 0.130615234375, 0.0989990234375, -1.091796875, -0.040771484375, 0.30517578125, -0.0247802734375, 0.26416015625, -0.08984375, 0.09765625, 0.15478515625, 0.253173828125, 0.55908203125, 0.30224609375, 0.4296875, -0.443603515625, -0.0048828125, 0.047637939453125, -0.00726318359375, 0.128173828125, 0.0130615234375, -0.0709228515625, 0.005859375, 0.701171875, -0.1787109375, -0.3193359375, -0.101318359375, -0.0987548828125, 0.1024169921875, -0.0792236328125, 0.081787109375, 0.2021484375, 0.0455322265625, 0.128662109375, 0.29931640625, -0.225341796875, -0.2081298828125, -0.155029296875, -0.38916015625, -0.6396484375, -0.0576171875, 0.6962890625, 0.1513671875, -0.252197265625, 0.05963134765625, 0.25146484375, 1.23828125, 0.193603515625, 0.2171630859375, 0.1064453125, 0.099853515625, -0.38818359375, -0.459716796875, 0.002685546875, -0.031982421875, 0.386962890625, 0.28466796875, 0.06494140625, 0.0042724609375, 0.43359375, 0.015869140625, 0.662109375, 0.1270751953125, 0.39453125, 0.54638671875, -0.1854248046875, 0.5361328125, -1.5283203125, -0.366943359375, -0.279541015625, 0.155029296875, -0.052490234375, -0.271728515625, 0.7412109375, 0.3173828125, -0.0609130859375, 0.031494140625, 0.292236328125, -0.08642578125, -0.27587890625, -0.974609375, -0.442626953125, -0.294921875, -0.402099609375, -0.869140625, 0.357177734375, 0.25537109375, 0.487548828125, -0.3974609375, -0.1298828125, 0.1058349609375, -0.078125, 0.3720703125, -0.2265625, 0.136962890625, 0.2027587890625, 0.458740234375, -0.406005859375, 0.1754150390625, -0.301025390625, -0.46728515625, -0.346435546875, 0.1162109375, -0.192626953125, -0.125, 0.009033203125, -0.10626220703125, 0.0068359375, 0.47802734375, -0.28662109375, 0.2274169921875, 0.404296875, 0.0333251953125, -0.2529296875, 0.02276611328125, 0.08489990234375, 0.0833740234375, -0.124267578125, -0.2724609375, -1.521484375, -0.48291015625, 0.218994140625, -0.015869140625, -0.337890625, -0.447509765625, -0.2044677734375, -0.090576171875, -0.059234619140625, -0.261962890625, -0.1884765625, -0.85107421875, 0.25048828125, 0.34619140625, -0.2432861328125, -0.2080078125, -0.1553955078125, -0.59228515625, 0.2308349609375, 0.3017578125, -0.00286865234375, -0.1783447265625, 0.05078125, -0.1328125, 0.0587158203125, 0.084228515625, -0.303955078125, -0.0103912353515625, 0.02935791015625, 0.43359375, 0.21240234375, -0.0682373046875, 0.1787109375, -0.01800537109375, -0.41162109375, 0.68115234375, 0.0994873046875, -0.159423828125, -0.3037109375, 1.357421875, 0.07470703125, -0.291015625, 0.054931640625, -0.393798828125, 0.35498046875, -0.244140625, -0.175537109375, 0.326904296875, 0.1304931640625, 0.0252685546875, 0.251220703125, 0.09417724609375, 0.26416015625, -0.281982421875, 0.359375, 0.250732421875, -0.64697265625, 0.14794921875, 0.278076171875, -0.1004638671875, -0.23876953125, -0.2301025390625, -0.301513671875, -0.8583984375, -0.01678466796875, -0.0638427734375, -0.08770751953125, -0.3623046875, 0.060546875, -0.0830078125, 0.43603515625, -0.446533203125, -1.365234375, -0.05029296875, -0.142822265625, 0.204833984375, 0.211669921875, -0.069091796875, 0.357666015625, 0.1051025390625, -0.078369140625, 0.335693359375, -0.007568359375, 0.251953125, 0.474853515625, 0.0625, 0.130859375, -0.157470703125, 0.31396484375, 0.17822265625, -0.4013671875, -0.61767578125, -0.1778564453125, -0.37841796875, -0.02294921875, 0.2705078125, -0.267578125, 0.87548828125, -0.093017578125, 0.1036376953125, -0.87890625, 0.0009765625, -0.06146240234375, 0.1312255859375, -0.5888671875, 0.265869140625, -0.58203125, 0.38330078125, -0.032745361328125, 0.2822265625, -0.131103515625, -0.51416015625, -0.6689453125, 0.1094970703125, -0.486328125, 0.496826171875, -0.2490234375, 0.990234375, -0.322265625, 0.04071044921875, -0.144775390625, -0.0623779296875, -0.43994140625, -0.003662109375, 1.40625, 0.050994873046875, 0.308349609375, -0.0732421875, 0.407958984375, -1.109375, -0.1796875, 0.5029296875, 0.0384521484375, -0.2310791015625, 0.59130859375, 0.147216796875, 0.101318359375, 0.200439453125, 0.06951904296875, 0.16015625, -0.07470703125, 0.51171875, -0.5478515625, 0.046142578125, -0.5693359375, -0.003173828125, 0.1121826171875, -0.49755859375, -0.02288818359375, -0.0511474609375, -0.1038818359375, 0.0628662109375, -0.8486328125, -0.186279296875, 0.2646484375, -0.07403564453125, 0.2548828125, 0.00933837890625, -0.107177734375, -0.7939453125, 0.027099609375, 0.279541015625, -0.02642822265625, 0.09375, -0.238525390625, -0.5673828125, 0.56787109375, -0.453857421875, 0.0361328125, -0.6044921875, 0.067626953125, -0.44091796875, 0.058013916015625, -0.0263671875, 0.07403564453125, 0.0278472900390625, 0.2222900390625, -2.25390625, 0.28271484375, -0.171875, 0.251220703125, 0.6865234375, -0.0166015625, -0.330322265625, -0.1806640625, 0.24072265625, -0.147705078125, 0.21142578125, 0.05419921875, 0.072509765625, 0.5244140625, 0.3876953125, 0.34033203125, -0.0118408203125, 0.2548828125, 0.63671875, -0.28466796875, 0.3466796875, 0.0447998046875, 0.04974365234375, -0.302490234375, 0.07666015625, 0.2305908203125, 0.3505859375, -0.7607421875, 0.603515625, 0.091796875, -0.2137451171875, 0.051513671875, -0.45166015625, -0.0032958984375, -0.00262451171875, 0.339599609375, -0.251953125, -0.016021728515625, -0.18896484375, -0.17919921875, 0.441650390625, -0.0081787109375, -0.32666015625, -0.6328125, -0.3544921875, 0.25439453125, 0.0924072265625, -0.20556640625, 0.33251953125, -0.091064453125, 0.276123046875, -0.419189453125, -0.2646484375, 0.197998046875, -0.3994140625, -0.45556640625, -0.6171875, -0.57177734375, 0.0081787109375, 0.7763671875, 0.66259765625, -0.264892578125, 0.03564453125, 0.6552734375, -0.6171875, 0.261474609375, -0.6416015625, -0.736328125, -0.48681640625, 0.15869140625, 0.0721435546875, -0.3544921875, -0.2998046875, 0.2259521484375, -0.10888671875, 0.16015625, 0.404541015625, 0.0731201171875, 0.0906982421875, 0.014251708984375, -0.5224609375, 0.104736328125, -0.054443359375, -0.04541015625, 0.755859375, 0.05126953125, -0.312255859375, 0.466064453125, -0.6337890625, 0.451171875, -0.2763671875, 0.5966796875, -0.1373291015625, 0.118896484375, 0.1661376953125, -0.96484375, -0.126953125, 0.014892578125, 0.193115234375, -0.29150390625, -0.43701171875, -0.0242919921875, 1.056640625, 0.368896484375, -0.62451171875, -0.225341796875, 0.6650390625, -0.1314697265625, -0.1728515625, -0.1007080078125, -0.427734375, 0.261962890625, -0.153076171875, -0.452392578125, 0.449951171875, 0.41845703125, -0.28173828125, -0.5634765625, 0.09222412109375, 0.0706787109375, -0.1180419921875, 0.1446533203125, -0.243408203125, -0.07318115234375, 0.175048828125, 0.3681640625, 0.375, 0.376708984375, 0.2197265625, -0.099853515625, 0.15869140625, -0.03924560546875, 0.36669921875, 0.33544921875, 0.4921875, -0.337890625, -0.10614013671875, 0.169677734375, -0.31591796875, -0.021484375, -0.217041015625, 0.36279296875, 0.411865234375, -0.239990234375, 0.498779296875, -2.123046875, 0.53125, 0.413330078125, 0.100341796875, -0.436279296875, -0.1539306640625, 0.0888671875, -0.140380859375, 0.0767822265625, 0.2607421875, 0.392822265625, -0.162841796875, 0.80712890625, -0.3603515625, 0.0501708984375, 0.69091796875, 0.2388916015625, 0.2529296875, -0.2001953125, -0.39990234375, 0.2130126953125, 0.4755859375, 0.1971435546875, -0.09442138671875, -0.236572265625, 0.2919921875, 0.131103515625, 0.9443359375, 0.259033203125, 0.3828125, -0.31201171875, 0.384033203125, 0.18115234375, -0.0045166015625, 0.6884765625, -0.69384765625, -0.320556640625, 0.51708984375, -0.6142578125, -0.0052490234375, 0.767578125, -0.4833984375, -0.228515625, -0.0511474609375, -0.41943359375, -0.232666015625, 0.177001953125, -0.9775390625, -0.102783203125, 0.10467529296875, 0.00146484375, 0.6220703125, -0.2744140625, 0.41552734375, -0.295166015625, -0.22607421875, -0.2230224609375, -0.052978515625, 0.085693359375, -0.0716552734375, 0.10693359375, 0.00998687744140625, 0.05657958984375, 0.30810546875, -0.0791015625, -0.133544921875, -0.341064453125, -0.003841400146484375, 0.49658203125, 1.2509765625, -0.15478515625, -0.222900390625, -0.5947265625, 0.541015625, 0.3466796875, 0.0552978515625, 0.30419921875, -0.263671875, -1.28125, -0.04901123046875, 0.0389404296875, -0.619140625, 0.451416015625, 0.5361328125, -0.0008544921875, 0.2587890625, -0.14306640625, 0.1175537109375, 0.1181640625, 0.01373291015625, -0.517578125, 0.3251953125, -0.15380859375, -0.48876953125, 0.20703125, -0.139404296875, -0.29443359375, -0.195556640625, -0.1461181640625, -0.192138671875, -0.397705078125, 0.052734375, -0.01220703125, -0.384765625, -0.53125, -0.12841796875, -0.5556640625, -0.10394287109375, 0.2724609375, -0.09228515625, -0.28466796875, -0.181396484375, -0.197998046875, -0.1627197265625, 0.216064453125, 0.54248046875, 0.1005859375, 0.21240234375, -0.708984375, -0.317138671875, -0.476806640625, -0.261962890625, 0.0052490234375, 0.1702880859375, 0.222412109375, 0.036376953125, 0.401611328125, -0.048614501953125, 0.16943359375, -0.0565185546875, -0.2384033203125, 0.424560546875, 0.158935546875, 0.57080078125, -0.1142578125, 0.05908203125, -0.79052734375, -0.06719970703125, 0.0643310546875, 0.1800537109375, -0.26416015625, -0.0770263671875, -0.533203125, 0.2880859375, -1.4248046875, 0.1513671875, -0.139404296875, 1.2333984375, -0.107421875, 0.37744140625, 0.37353515625, 0.381103515625, -0.201171875, -0.0400390625, 0.263916015625, -0.0478515625, 0.40234375, 0.55859375, -0.61279296875, -0.234130859375, -0.20703125, -0.12744140625, -0.0626220703125, 0.183349609375, 0.43212890625, 0.0819091796875, -0.55908203125, 0.22509765625, -0.214111328125, -0.048553466796875, -0.70166015625, 0.884765625, -0.03564453125, -0.0184326171875, -0.001953125, -0.728515625, 0.473876953125, 0.299072265625, 0.52685546875, -0.1151123046875, -0.7177734375, 0.66748046875, 0.315185546875, -0.06414794921875, 1.0927734375, 0.034912109375, -0.1077880859375, -0.7705078125, -0.1121826171875, -0.224853515625, -0.0072021484375, 0.024169921875, 0.1380615234375, 0.439453125, -0.18798828125, 0.0660400390625, -0.017822265625, -0.5, -0.30224609375, 0.29296875, -0.484375, -0.317138671875, 0.1287841796875, -0.362060546875, -0.077392578125, -0.160888671875, -0.02783203125, 0.8740234375, 0.0010986328125, -0.266845703125, -0.139892578125, 0.1695556640625, 0.110107421875, -0.0328369140625, 0.2353515625, -0.114990234375, 0.2408447265625, 0.57373046875, -0.390625, 0.1513671875, -0.322021484375, -0.826171875, 0.1260986328125, 0.2548828125, 0.1455078125, 0.0103759765625, 0.0770263671875, 0.0830078125, 0.1114501953125, -0.198486328125, 0.40087890625, 0.67431640625]}, {"frame_number": 20, "vector": [-0.2218017578125, 0.25830078125, 0.012451171875, -0.1427001953125, -0.6455078125, 0.329833984375, 0.0372314453125, -0.454345703125, -0.09130859375, -0.08258056640625, -0.1123046875, 0.021484375, 0.31494140625, 0.2353515625, 0.02099609375, 0.2626953125, 0.090576171875, -0.142822265625, -0.08026123046875, -0.4716796875, -0.59375, -6.1796875, -0.35009765625, -0.66259765625, 0.0684814453125, -0.1083984375, -0.26123046875, -0.849609375, 0.489013671875, -0.18798828125, -0.90869140625, -0.6416015625, -0.634765625, 0.06622314453125, -0.06842041015625, 0.0784912109375, 0.355712890625, -0.62890625, -0.04046630859375, -0.45947265625, -0.03466796875, 0.238037109375, 0.154296875, 0.016845703125, 1.0263671875, -0.36083984375, -0.39892578125, -0.177490234375, -0.57177734375, -0.11572265625, 0.019256591796875, -0.294677734375, 0.0904541015625, 0.484375, 0.232421875, 0.01136016845703125, 0.38037109375, -0.277099609375, 0.2822265625, 0.0543212890625, -0.064208984375, 0.156982421875, 0.265869140625, 0.1680908203125, -0.25, -0.32666015625, -0.47802734375, 0.01708984375, 0.236083984375, 0.15380859375, 0.953125, -0.03369140625, 0.025634765625, -0.0364990234375, -0.362548828125, 0.82373046875, -0.299072265625, -0.41552734375, 0.1708984375, -0.70458984375, 1.162109375, 0.307861328125, 0.61328125, 0.5224609375, -0.2666015625, 0.08563232421875, 0.226318359375, 0.00262451171875, 0.177734375, -0.62353515625, -0.10540771484375, -0.10009765625, 0.0048828125, -0.11083984375, 0.284912109375, 0.34814453125, 0.077392578125, 0.51025390625, 0.019805908203125, -0.65869140625, 0.69677734375, 0.35400390625, -0.39599609375, -0.65576171875, 0.040771484375, -0.0697021484375, 0.0782470703125, -0.115478515625, -0.146484375, -0.021240234375, -0.0821533203125, 0.133056640625, 0.09722900390625, -0.419677734375, 0.13671875, -0.548828125, -0.171875, -0.123046875, 0.314453125, 0.52978515625, 0.066162109375, -0.269775390625, -0.05963134765625, -0.62890625, -0.0340576171875, 0.33349609375, -0.16845703125, 0.3740234375, 0.1732177734375, 0.239013671875, 0.0008544921875, -0.463623046875, -0.0543212890625, -0.0518798828125, 0.52783203125, -0.0399169921875, 0.0850830078125, 0.75830078125, 0.5791015625, 0.04541015625, 0.00140380859375, 0.349609375, 0.149658203125, -0.1204833984375, -0.289306640625, 0.128662109375, -0.026397705078125, -0.407958984375, 0.235595703125, 1.109375, -0.150390625, -1.5712890625, 0.1285400390625, -0.48681640625, 0.41162109375, -0.26953125, -0.19140625, 0.994140625, 0.352783203125, -0.4384765625, -0.26953125, 0.215576171875, 0.400390625, 0.40625, -0.27099609375, -0.261962890625, -0.01025390625, 0.6552734375, 0.080078125, 0.0286865234375, 0.2178955078125, 0.443115234375, -0.184326171875, 0.583984375, -0.296142578125, 0.947265625, -1.0654296875, 0.13720703125, 0.037109375, -0.0712890625, 0.0634765625, 0.16357421875, -0.230224609375, 0.07110595703125, -0.0020751953125, 0.025146484375, -0.032470703125, 0.1512451171875, -0.0731201171875, 0.12109375, 0.19091796875, 0.1783447265625, 0.48974609375, -0.39306640625, 0.53564453125, -0.215087890625, 0.42138671875, -0.25537109375, -0.30908203125, 1.1484375, 0.5146484375, -0.02099609375, -0.1962890625, 0.0380859375, -0.04901123046875, 0.110595703125, 0.65478515625, 0.01287841796875, -0.615234375, -0.1683349609375, 0.1790771484375, -0.082763671875, 0.2427978515625, -0.8505859375, -0.5361328125, -0.56201171875, 0.50634765625, 0.98486328125, -0.0248260498046875, -0.42529296875, -0.437744140625, -0.48974609375, -0.143798828125, -0.60595703125, -0.146728515625, 0.216796875, -0.098388671875, -0.1953125, 0.67919921875, 0.03369140625, -0.06756591796875, 0.28857421875, 0.0001220703125, 0.3505859375, -0.348876953125, -0.59912109375, 0.73779296875, -0.0748291015625, -0.97412109375, -0.0810546875, 0.71142578125, 0.2646484375, -0.0562744140625, 0.216552734375, -0.158447265625, 1.0390625, -0.19775390625, -0.00537109375, -0.29296875, -0.135009765625, 0.0445556640625, -0.0355224609375, 0.328857421875, -0.060302734375, -0.6669921875, -0.02880859375, -0.1143798828125, -0.0478515625, 0.447998046875, 0.466552734375, -0.047332763671875, -0.2015380859375, -0.057861328125, -0.0540771484375, -0.271240234375, -0.031494140625, -0.826171875, -0.33642578125, 0.55859375, -0.36767578125, 0.00067138671875, 0.44189453125, 0.35595703125, 0.2509765625, -0.2352294921875, -0.69580078125, -0.169189453125, 0.15234375, -0.3544921875, -0.42041015625, -0.283935546875, 0.67529296875, -0.39404296875, 0.17626953125, 0.400634765625, 0.29345703125, -0.1468505859375, 0.541015625, -0.57958984375, 1.1923828125, -0.08331298828125, 0.125732421875, 0.022705078125, -0.444580078125, 0.06854248046875, 0.1053466796875, 0.124755859375, 0.411376953125, 0.568359375, -0.2607421875, 0.023590087890625, 0.748046875, -0.06219482421875, -0.038330078125, -0.34423828125, 0.4140625, 0.359375, -0.6640625, 0.12286376953125, -0.0816650390625, 0.32568359375, -0.3955078125, 0.00927734375, -0.2802734375, -0.32861328125, 0.35693359375, 0.00439453125, 0.1011962890625, -0.26953125, 0.119140625, -0.90625, 0.182373046875, 0.15380859375, -0.44873046875, -0.89892578125, 0.54443359375, -0.43212890625, 0.25390625, -0.5927734375, -0.06695556640625, -0.5341796875, -0.0765380859375, -0.123046875, -0.471923828125, 0.1810302734375, 0.310302734375, 0.126953125, -0.6494140625, 0.054290771484375, -0.015869140625, -0.73876953125, 0.1337890625, -0.292724609375, -0.69873046875, -0.587890625, 0.203857421875, -0.41455078125, 0.368896484375, 0.1932373046875, 0.001953125, -0.336669921875, -0.31298828125, 0.317138671875, 0.178466796875, 0.392333984375, 0.32470703125, 0.2474365234375, -0.16162109375, -0.29248046875, -0.2254638671875, -0.214111328125, -0.0321044921875, 0.5087890625, 0.256103515625, -0.62890625, -0.455078125, 0.0068359375, 0.2362060546875, 0.02880859375, -0.12176513671875, -0.009765625, 0.218505859375, -0.100830078125, 0.37158203125, 0.35205078125, -0.112060546875, 0.42724609375, -0.281982421875, 0.07421875, -0.58935546875, -0.06005859375, -0.28759765625, 0.093994140625, 0.74658203125, 0.1798095703125, -0.0213623046875, -0.05633544921875, -0.0927734375, -0.25048828125, 0.51953125, -0.65234375, -0.156005859375, 0.265869140625, 0.279541015625, 1.005859375, 0.63525390625, -0.106689453125, -0.69482421875, -0.0577392578125, -0.30712890625, 0.20458984375, 0.223876953125, -0.1182861328125, -0.36767578125, -0.169189453125, 0.0457763671875, -0.3740234375, 0.05572509765625, 0.36865234375, -0.120849609375, 0.25341796875, -0.89990234375, -0.7021484375, -0.29296875, -0.3740234375, 0.5009765625, 0.7587890625, -0.27978515625, 1.087890625, -0.2958984375, 0.160400390625, -0.06549072265625, -0.25537109375, -0.0941162109375, 0.0396728515625, -0.490234375, -1.3388671875, -0.0699462890625, 0.1478271484375, 0.2169189453125, 0.084228515625, -6.1875, -0.38330078125, -0.056884765625, -0.74169921875, -0.263671875, -0.1409912109375, -1.19921875, -0.6708984375, -0.2626953125, -0.0087890625, 0.6298828125, 0.62353515625, -0.2451171875, -0.009033203125, 0.2734375, 0.1754150390625, 0.0135498046875, -0.143798828125, 0.170654296875, -0.041748046875, 0.201416015625, -0.09503173828125, 0.1617431640625, 0.416259765625, -0.251953125, -0.1578369140625, 0.321533203125, -0.60546875, -0.6162109375, -0.13037109375, 0.101318359375, -0.3212890625, -0.28125, -0.46875, -0.0533447265625, 0.57666015625, 0.7900390625, -0.27978515625, -0.46875, 0.2415771484375, -0.97314453125, 0.328369140625, -0.029296875, -0.07647705078125, 0.211181640625, 0.334228515625, -0.0994873046875, -0.744140625, -0.78076171875, 0.1209716796875, 0.06103515625, -1.1025390625, -0.07568359375, 0.3046875, -0.0487060546875, 0.239990234375, -0.08856201171875, 0.0845947265625, 0.1678466796875, 0.293212890625, 0.5615234375, 0.28857421875, 0.40478515625, -0.4296875, 0.02783203125, 0.06292724609375, 0.0029296875, 0.10107421875, 0.017333984375, -0.095458984375, 0.005157470703125, 0.68798828125, -0.175537109375, -0.36572265625, -0.08642578125, -0.073486328125, 0.0928955078125, -0.1026611328125, 0.124755859375, 0.208984375, 0.0467529296875, 0.126220703125, 0.275390625, -0.23681640625, -0.1959228515625, -0.107177734375, -0.381591796875, -0.677734375, -0.0860595703125, 0.7275390625, 0.23193359375, -0.209228515625, 0.05413818359375, 0.2156982421875, 1.25, 0.18896484375, 0.2352294921875, 0.1107177734375, 0.1287841796875, -0.4013671875, -0.44775390625, 0.055908203125, -0.03564453125, 0.38525390625, 0.264892578125, 0.048095703125, 0.00286865234375, 0.43798828125, 0.0350341796875, 0.6494140625, 0.11279296875, 0.404296875, 0.5263671875, -0.2083740234375, 0.54638671875, -1.525390625, -0.34423828125, -0.275390625, 0.150634765625, -0.03466796875, -0.2822265625, 0.74072265625, 0.2880859375, -0.05181884765625, 0.049072265625, 0.350341796875, -0.10302734375, -0.27294921875, -0.97216796875, -0.465087890625, -0.320556640625, -0.39306640625, -0.8427734375, 0.36572265625, 0.2275390625, 0.48828125, -0.41650390625, -0.1380615234375, 0.0936279296875, -0.07421875, 0.36669921875, -0.26953125, 0.09649658203125, 0.2093505859375, 0.430908203125, -0.4296875, 0.1878662109375, -0.3095703125, -0.461181640625, -0.32177734375, 0.089599609375, -0.170654296875, -0.1365966796875, 0.0101318359375, -0.07232666015625, 0.03369140625, 0.45458984375, -0.27587890625, 0.25537109375, 0.4296875, 0.0234527587890625, -0.1932373046875, 0.032867431640625, 0.076416015625, 0.1109619140625, -0.12451171875, -0.263427734375, -1.505859375, -0.45849609375, 0.22509765625, -0.04541015625, -0.353271484375, -0.420654296875, -0.18115234375, -0.093994140625, -0.0247650146484375, -0.2091064453125, -0.2205810546875, -0.87890625, 0.19921875, 0.328125, -0.248291015625, -0.212158203125, -0.1668701171875, -0.53466796875, 0.2178955078125, 0.304443359375, 0.0439453125, -0.209228515625, 0.0751953125, -0.11474609375, 0.04217529296875, 0.08245849609375, -0.32080078125, -0.021820068359375, 0.01617431640625, 0.349365234375, 0.21533203125, -0.0836181640625, 0.209716796875, 0.00531005859375, -0.40966796875, 0.69580078125, 0.0936279296875, -0.145751953125, -0.3115234375, 1.337890625, 0.107666015625, -0.30517578125, 0.0914306640625, -0.408203125, 0.35302734375, -0.20751953125, -0.154052734375, 0.315673828125, 0.14208984375, 0.029052734375, 0.28515625, 0.0599365234375, 0.2452392578125, -0.280029296875, 0.337890625, 0.27490234375, -0.65185546875, 0.174560546875, 0.271728515625, -0.13916015625, -0.2412109375, -0.22216796875, -0.32080078125, -0.8212890625, -0.009521484375, -0.055419921875, -0.08575439453125, -0.3515625, 0.083740234375, -0.10205078125, 0.4296875, -0.4111328125, -1.40234375, -0.075927734375, -0.0919189453125, 0.178955078125, 0.184326171875, -0.06195068359375, 0.35302734375, 0.0859375, -0.062744140625, 0.3125, 0.0133056640625, 0.235107421875, 0.4638671875, 0.066162109375, 0.11669921875, -0.17138671875, 0.3037109375, 0.1494140625, -0.397705078125, -0.6259765625, -0.18115234375, -0.337890625, 0.04541015625, 0.232177734375, -0.259033203125, 0.82177734375, -0.0889892578125, 0.11273193359375, -0.892578125, 0.03289794921875, -0.02410888671875, 0.137451171875, -0.5966796875, 0.2861328125, -0.5830078125, 0.35546875, -0.050048828125, 0.25634765625, -0.107177734375, -0.48779296875, -0.6484375, 0.06170654296875, -0.465576171875, 0.48193359375, -0.251220703125, 0.962890625, -0.28076171875, 0.0845947265625, -0.1416015625, -0.07763671875, -0.3974609375, 0.025146484375, 1.41796875, 0.05999755859375, 0.32421875, -0.0712890625, 0.383544921875, -1.0703125, -0.16650390625, 0.53271484375, 0.023681640625, -0.22265625, 0.517578125, 0.0933837890625, 0.087890625, 0.247314453125, 0.0191650390625, 0.180419921875, -0.0732421875, 0.49267578125, -0.5771484375, 0.088623046875, -0.5791015625, 0.0338134765625, 0.0816650390625, -0.48388671875, -0.04083251953125, -0.027099609375, -0.09637451171875, 0.0999755859375, -0.8173828125, -0.1903076171875, 0.30419921875, -0.12054443359375, 0.2685546875, 0.01885986328125, -0.1011962890625, -0.8310546875, 0.0264892578125, 0.253173828125, 0.03466796875, 0.1407470703125, -0.25048828125, -0.57177734375, 0.5625, -0.443603515625, 0.057861328125, -0.583984375, 0.0789794921875, -0.423828125, 0.076904296875, 0.025390625, 0.0693359375, 0.0128936767578125, 0.2330322265625, -2.259765625, 0.2646484375, -0.133056640625, 0.265869140625, 0.64599609375, 0.00146484375, -0.33203125, -0.1875, 0.15478515625, -0.1331787109375, 0.220947265625, 0.050048828125, 0.126708984375, 0.492431640625, 0.338134765625, 0.3388671875, 0.0103759765625, 0.31103515625, 0.6005859375, -0.2763671875, 0.3623046875, 0.066650390625, 0.01806640625, -0.329833984375, 0.081298828125, 0.235107421875, 0.34814453125, -0.69580078125, 0.59375, 0.080810546875, -0.1729736328125, 0.030517578125, -0.42724609375, 0.001953125, -0.001708984375, 0.37939453125, -0.296875, 0.020263671875, -0.184326171875, -0.187255859375, 0.412109375, -0.029876708984375, -0.3076171875, -0.6552734375, -0.333740234375, 0.254150390625, 0.0758056640625, -0.2047119140625, 0.310302734375, -0.0767822265625, 0.29833984375, -0.432373046875, -0.316650390625, 0.235595703125, -0.365966796875, -0.411376953125, -0.62255859375, -0.62060546875, 0.01666259765625, 0.7978515625, 0.68115234375, -0.3212890625, 0.028076171875, 0.63525390625, -0.59130859375, 0.21728515625, -0.71435546875, -0.712890625, -0.49853515625, 0.165283203125, 0.0653076171875, -0.356689453125, -0.29052734375, 0.170166015625, -0.13525390625, 0.16357421875, 0.43212890625, 0.10479736328125, 0.0938720703125, 0.024383544921875, -0.5390625, 0.1015625, -0.0653076171875, -0.0628662109375, 0.78466796875, 0.099365234375, -0.360107421875, 0.46142578125, -0.591796875, 0.4423828125, -0.269775390625, 0.533203125, -0.12127685546875, 0.067626953125, 0.1585693359375, -0.9609375, -0.1259765625, 0.0465087890625, 0.2021484375, -0.267822265625, -0.4267578125, 0.0008544921875, 1.0263671875, 0.36767578125, -0.62939453125, -0.209716796875, 0.6982421875, -0.147216796875, -0.18505859375, -0.07635498046875, -0.4248046875, 0.236083984375, -0.15625, -0.45751953125, 0.436767578125, 0.433837890625, -0.297119140625, -0.5732421875, 0.0714111328125, 0.08984375, -0.11346435546875, 0.158447265625, -0.25244140625, -0.09210205078125, 0.23974609375, 0.340087890625, 0.35498046875, 0.36572265625, 0.18994140625, -0.093017578125, 0.187744140625, -0.023193359375, 0.359375, 0.343994140625, 0.48779296875, -0.318359375, -0.080078125, 0.14990234375, -0.29541015625, -0.1142578125, -0.25439453125, 0.364013671875, 0.444091796875, -0.255615234375, 0.466552734375, -2.11328125, 0.49951171875, 0.43994140625, 0.15576171875, -0.462890625, -0.178466796875, 0.10107421875, -0.1748046875, 0.06475830078125, 0.22021484375, 0.42041015625, -0.174560546875, 0.74755859375, -0.352783203125, 0.07220458984375, 0.65087890625, 0.2452392578125, 0.2509765625, -0.2353515625, -0.37158203125, 0.1807861328125, 0.4462890625, 0.17138671875, -0.06292724609375, -0.30322265625, 0.29296875, 0.142578125, 0.94873046875, 0.26171875, 0.35498046875, -0.299560546875, 0.369384765625, 0.182373046875, -0.00531005859375, 0.7861328125, -0.6640625, -0.295166015625, 0.474609375, -0.61181640625, -0.028076171875, 0.7724609375, -0.4560546875, -0.2314453125, -0.13232421875, -0.43408203125, -0.122314453125, 0.166748046875, -0.97119140625, -0.0906982421875, 0.0867919921875, -0.03515625, 0.6357421875, -0.2939453125, 0.443115234375, -0.2744140625, -0.2384033203125, -0.2342529296875, -0.04931640625, 0.0703125, -0.1339111328125, 0.1444091796875, -0.013885498046875, 0.0860595703125, 0.28515625, -0.06005859375, -0.117431640625, -0.359375, 0.0141754150390625, 0.44677734375, 1.2880859375, -0.15283203125, -0.220458984375, -0.56005859375, 0.5439453125, 0.32470703125, 0.0487060546875, 0.31982421875, -0.2783203125, -1.365234375, -0.04632568359375, 0.0665283203125, -0.6005859375, 0.4052734375, 0.5771484375, 0.00225830078125, 0.225830078125, -0.1290283203125, 0.092529296875, 0.124267578125, 0.04437255859375, -0.5654296875, 0.338134765625, -0.1087646484375, -0.509765625, 0.272216796875, -0.0906982421875, -0.314453125, -0.185302734375, -0.135986328125, -0.14990234375, -0.44921875, 0.04150390625, 0.0140380859375, -0.395263671875, -0.5, -0.15380859375, -0.50244140625, -0.11328125, 0.3173828125, -0.054443359375, -0.251953125, -0.19287109375, -0.1875, -0.163330078125, 0.15185546875, 0.49462890625, 0.04913330078125, 0.176025390625, -0.69189453125, -0.31689453125, -0.51318359375, -0.24658203125, -0.0050048828125, 0.205810546875, 0.2213134765625, 0.03009033203125, 0.3916015625, -0.0129852294921875, 0.1279296875, -0.02276611328125, -0.275390625, 0.424072265625, 0.122314453125, 0.56787109375, -0.117919921875, 0.04150390625, -0.74755859375, -0.06512451171875, 0.0140380859375, 0.1661376953125, -0.246826171875, -0.0821533203125, -0.5283203125, 0.28955078125, -1.43359375, 0.12109375, -0.128662109375, 1.2314453125, -0.15380859375, 0.414306640625, 0.3662109375, 0.34375, -0.20361328125, -0.02374267578125, 0.26904296875, -0.05029296875, 0.373291015625, 0.5703125, -0.576171875, -0.212158203125, -0.182373046875, -0.125732421875, -0.00115966796875, 0.19140625, 0.46044921875, 0.05047607421875, -0.5869140625, 0.2215576171875, -0.2216796875, -0.048492431640625, -0.703125, 0.84619140625, -0.00830078125, -0.004638671875, -0.0098876953125, -0.7099609375, 0.463623046875, 0.312744140625, 0.52197265625, -0.128662109375, -0.73046875, 0.5791015625, 0.378662109375, -0.083984375, 1.0927734375, -0.0303955078125, -0.094970703125, -0.755859375, -0.1348876953125, -0.197265625, -0.0572509765625, 0.014404296875, 0.1768798828125, 0.46826171875, -0.174560546875, 0.11370849609375, 0.0228271484375, -0.52490234375, -0.28271484375, 0.225830078125, -0.4462890625, -0.302978515625, 0.1043701171875, -0.36865234375, -0.05322265625, -0.126708984375, 0.017822265625, 0.8486328125, 0.048583984375, -0.26708984375, -0.154541015625, 0.178955078125, 0.07220458984375, -0.00274658203125, 0.248046875, -0.13525390625, 0.25927734375, 0.5537109375, -0.37939453125, 0.19384765625, -0.30078125, -0.794921875, 0.1412353515625, 0.24462890625, 0.153076171875, 0.0335693359375, 0.0748291015625, 0.0654296875, 0.0677490234375, -0.21337890625, 0.404541015625, 0.69677734375]}, {"frame_number": 21, "vector": [-0.235595703125, 0.2548828125, 0.019775390625, -0.1470947265625, -0.5830078125, 0.33447265625, 0.0745849609375, -0.447021484375, -0.09521484375, -0.09185791015625, -0.112060546875, 0.03173828125, 0.28466796875, 0.253662109375, 0.0211181640625, 0.300537109375, 0.086181640625, -0.1865234375, -0.06787109375, -0.49755859375, -0.5869140625, -6.09375, -0.353515625, -0.69287109375, 0.05767822265625, -0.10980224609375, -0.25, -0.87744140625, 0.51171875, -0.180908203125, -0.91748046875, -0.62939453125, -0.6591796875, 0.0616455078125, -0.1011962890625, 0.039306640625, 0.35400390625, -0.6533203125, -0.0762939453125, -0.454345703125, -0.04443359375, 0.1990966796875, 0.130615234375, 0.048583984375, 1.021484375, -0.37109375, -0.38134765625, -0.1712646484375, -0.57421875, -0.13134765625, -0.0174102783203125, -0.280029296875, 0.10888671875, 0.45849609375, 0.2294921875, -0.02716064453125, 0.352783203125, -0.280517578125, 0.271484375, 0.0306396484375, -0.033203125, 0.1363525390625, 0.266845703125, 0.1561279296875, -0.2340087890625, -0.298828125, -0.485107421875, -0.017822265625, 0.261474609375, 0.1292724609375, 0.9384765625, -0.031005859375, 0.0126953125, -0.0400390625, -0.380126953125, 0.82421875, -0.31103515625, -0.41748046875, 0.194580078125, -0.6787109375, 1.10546875, 0.3125, 0.587890625, 0.5, -0.25244140625, 0.1181640625, 0.215576171875, 0.01885986328125, 0.17724609375, -0.6298828125, -0.09088134765625, -0.1138916015625, 0.0126953125, -0.089111328125, 0.281494140625, 0.336181640625, 0.098876953125, 0.498046875, 0.03582763671875, -0.64306640625, 0.6982421875, 0.349609375, -0.41259765625, -0.66650390625, 0.03564453125, -0.0875244140625, 0.06707763671875, -0.1358642578125, -0.1553955078125, 0.010498046875, -0.08978271484375, 0.140869140625, 0.0677490234375, -0.37255859375, 0.121337890625, -0.54296875, -0.09619140625, -0.10888671875, 0.341796875, 0.5244140625, 0.0732421875, -0.26171875, -0.0631103515625, -0.6025390625, -0.011962890625, 0.340087890625, -0.174560546875, 0.39990234375, 0.1676025390625, 0.2587890625, -0.004638671875, -0.456298828125, -0.10693359375, -0.040008544921875, 0.492431640625, -0.04541015625, 0.1134033203125, 0.783203125, 0.55126953125, 0.072021484375, -0.054840087890625, 0.316650390625, 0.149658203125, -0.1337890625, -0.280517578125, 0.1409912109375, -0.013275146484375, -0.3974609375, 0.247314453125, 1.072265625, -0.13671875, -1.587890625, 0.1356201171875, -0.4912109375, 0.394287109375, -0.28564453125, -0.1536865234375, 1.0361328125, 0.371826171875, -0.427001953125, -0.269775390625, 0.195556640625, 0.396728515625, 0.402099609375, -0.291015625, -0.247802734375, -0.008056640625, 0.666015625, 0.0615234375, -0.03521728515625, 0.226318359375, 0.4296875, -0.1622314453125, 0.59228515625, -0.315185546875, 0.97412109375, -1.060546875, 0.1175537109375, 0.0509033203125, -0.06689453125, 0.0482177734375, 0.13720703125, -0.210205078125, 0.05694580078125, 0.0025634765625, 0.005859375, -0.046142578125, 0.162353515625, -0.0789794921875, 0.100341796875, 0.177490234375, 0.1553955078125, 0.48486328125, -0.3857421875, 0.515625, -0.2210693359375, 0.431640625, -0.26416015625, -0.2734375, 1.15625, 0.5166015625, -0.06201171875, -0.199951171875, 0.048828125, -0.0352783203125, 0.1190185546875, 0.68359375, 0.04266357421875, -0.6279296875, -0.1845703125, 0.170654296875, -0.12060546875, 0.2073974609375, -0.8828125, -0.5107421875, -0.5341796875, 0.490966796875, 1.037109375, -0.042022705078125, -0.446044921875, -0.426513671875, -0.474609375, -0.1221923828125, -0.58056640625, -0.16015625, 0.157470703125, -0.093994140625, -0.1865234375, 0.66796875, 0.06396484375, -0.064453125, 0.2705078125, -0.005126953125, 0.34423828125, -0.35400390625, -0.58203125, 0.705078125, -0.096923828125, -0.94189453125, -0.09814453125, 0.73046875, 0.280029296875, -0.054931640625, 0.248779296875, -0.1815185546875, 1.021484375, -0.19921875, 0.00146484375, -0.2274169921875, -0.160888671875, 0.0711669921875, -0.02825927734375, 0.3271484375, -0.072509765625, -0.69287109375, -0.030029296875, -0.09564208984375, -0.029296875, 0.4169921875, 0.46240234375, -0.0247802734375, -0.1917724609375, -0.0267333984375, -0.0472412109375, -0.28515625, -0.029296875, -0.8115234375, -0.307373046875, 0.54296875, -0.3505859375, -0.0010986328125, 0.46435546875, 0.34033203125, 0.23828125, -0.2196044921875, -0.7021484375, -0.1412353515625, 0.16259765625, -0.366455078125, -0.4189453125, -0.283935546875, 0.671875, -0.3720703125, 0.182373046875, 0.422119140625, 0.26318359375, -0.1412353515625, 0.5419921875, -0.57373046875, 1.15234375, -0.05291748046875, 0.143310546875, 0.0042724609375, -0.416259765625, 0.052978515625, 0.080810546875, 0.1240234375, 0.37841796875, 0.5498046875, -0.248779296875, 0.00911712646484375, 0.7421875, -0.055450439453125, -0.0458984375, -0.34619140625, 0.41259765625, 0.32568359375, -0.67822265625, 0.128173828125, -0.09033203125, 0.3193359375, -0.380126953125, 0.019775390625, -0.259033203125, -0.3193359375, 0.35693359375, 0.00738525390625, 0.086181640625, -0.2423095703125, 0.0751953125, -0.943359375, 0.1947021484375, 0.11859130859375, -0.436279296875, -0.8857421875, 0.5625, -0.406494140625, 0.23974609375, -0.55712890625, -0.07000732421875, -0.525390625, -0.067138671875, -0.1461181640625, -0.494140625, 0.177978515625, 0.3115234375, 0.12744140625, -0.681640625, 0.04913330078125, -0.050048828125, -0.7490234375, 0.161865234375, -0.27783203125, -0.681640625, -0.6142578125, 0.19580078125, -0.432861328125, 0.3876953125, 0.1868896484375, -0.001708984375, -0.321533203125, -0.294189453125, 0.290283203125, 0.18115234375, 0.411865234375, 0.310546875, 0.24462890625, -0.1455078125, -0.30078125, -0.236572265625, -0.23193359375, -0.0384521484375, 0.50244140625, 0.27685546875, -0.587890625, -0.456787109375, 0.000244140625, 0.228759765625, -0.007080078125, -0.1324462890625, -0.0074462890625, 0.2098388671875, -0.09521484375, 0.359375, 0.3603515625, -0.133056640625, 0.4033203125, -0.2783203125, 0.01318359375, -0.6435546875, -0.072265625, -0.266845703125, 0.1011962890625, 0.7529296875, 0.1552734375, -0.021484375, -0.0511474609375, -0.076171875, -0.23291015625, 0.5107421875, -0.6240234375, -0.14599609375, 0.264892578125, 0.271728515625, 1.0224609375, 0.67626953125, -0.141357421875, -0.6787109375, -0.0560302734375, -0.283203125, 0.20458984375, 0.2423095703125, -0.13232421875, -0.36376953125, -0.15771484375, 0.05712890625, -0.36767578125, 0.05755615234375, 0.37255859375, -0.1376953125, 0.246826171875, -0.87353515625, -0.68994140625, -0.27294921875, -0.351318359375, 0.4990234375, 0.7373046875, -0.277099609375, 1.123046875, -0.275390625, 0.1900634765625, -0.0887451171875, -0.257080078125, -0.075927734375, 0.0379638671875, -0.5537109375, -1.34765625, -0.0765380859375, 0.1748046875, 0.1893310546875, 0.058319091796875, -6.109375, -0.36572265625, -0.076171875, -0.7392578125, -0.2841796875, -0.0987548828125, -1.2099609375, -0.68310546875, -0.245849609375, -0.03106689453125, 0.640625, 0.59619140625, -0.236083984375, 0.0191650390625, 0.273681640625, 0.1763916015625, -0.0343017578125, -0.19140625, 0.197021484375, -0.029296875, 0.207275390625, -0.09039306640625, 0.1795654296875, 0.46142578125, -0.2381591796875, -0.14208984375, 0.322021484375, -0.6103515625, -0.634765625, -0.136962890625, 0.0830078125, -0.33447265625, -0.2822265625, -0.47412109375, -0.0784912109375, 0.609375, 0.78076171875, -0.290771484375, -0.443359375, 0.2440185546875, -0.98681640625, 0.31396484375, -0.0330810546875, -0.1134033203125, 0.19970703125, 0.35986328125, -0.087646484375, -0.78076171875, -0.791015625, 0.126953125, 0.0809326171875, -1.1103515625, -0.100830078125, 0.3037109375, -0.0592041015625, 0.270751953125, -0.105224609375, 0.05255126953125, 0.1900634765625, 0.243408203125, 0.529296875, 0.28955078125, 0.38671875, -0.437255859375, 0.033935546875, 0.052001953125, 0.0404052734375, 0.10302734375, 0.026611328125, -0.0836181640625, -0.02484130859375, 0.67529296875, -0.2001953125, -0.38037109375, -0.068359375, -0.060791015625, 0.0802001953125, -0.0604248046875, 0.127197265625, 0.2100830078125, 0.0518798828125, 0.11785888671875, 0.31396484375, -0.2412109375, -0.19580078125, -0.077392578125, -0.405029296875, -0.6806640625, -0.080810546875, 0.7177734375, 0.26025390625, -0.172119140625, 0.054107666015625, 0.22705078125, 1.2294921875, 0.1455078125, 0.249267578125, 0.0870361328125, 0.1259765625, -0.42138671875, -0.4580078125, 0.0654296875, -0.0244140625, 0.393798828125, 0.248779296875, 0.039306640625, -0.03070068359375, 0.40234375, 0.035888671875, 0.6220703125, 0.075439453125, 0.4111328125, 0.5107421875, -0.2078857421875, 0.52880859375, -1.5166015625, -0.35498046875, -0.2587890625, 0.1685791015625, -0.04248046875, -0.29443359375, 0.73828125, 0.291259765625, -0.04925537109375, 0.060791015625, 0.36279296875, -0.1005859375, -0.25244140625, -0.98193359375, -0.47802734375, -0.318603515625, -0.3896484375, -0.8701171875, 0.382080078125, 0.2052001953125, 0.5302734375, -0.40771484375, -0.1551513671875, 0.10614013671875, -0.0504150390625, 0.37353515625, -0.259765625, 0.12744140625, 0.220947265625, 0.468505859375, -0.44189453125, 0.2216796875, -0.33056640625, -0.51953125, -0.350341796875, 0.1107177734375, -0.185791015625, -0.1337890625, 0.00732421875, -0.021148681640625, 0.0302734375, 0.4677734375, -0.25, 0.255859375, 0.429443359375, 0.06939697265625, -0.1329345703125, 0.0565185546875, 0.10693359375, 0.0992431640625, -0.149658203125, -0.251953125, -1.5205078125, -0.45849609375, 0.23974609375, -0.05145263671875, -0.342529296875, -0.3955078125, -0.164794921875, -0.0791015625, -0.031097412109375, -0.195068359375, -0.22509765625, -0.888671875, 0.1785888671875, 0.341552734375, -0.20654296875, -0.217041015625, -0.151123046875, -0.52685546875, 0.2490234375, 0.287109375, 0.06689453125, -0.216552734375, 0.08056640625, -0.10498046875, 0.03594970703125, 0.0711669921875, -0.32568359375, -0.00213623046875, 0.00885009765625, 0.30419921875, 0.231201171875, -0.074462890625, 0.218994140625, -0.00360107421875, -0.39794921875, 0.6787109375, 0.0885009765625, -0.17578125, -0.3212890625, 1.3037109375, 0.125732421875, -0.29150390625, 0.0718994140625, -0.3984375, 0.3544921875, -0.215576171875, -0.148193359375, 0.31787109375, 0.11474609375, 0.0257568359375, 0.31005859375, 0.06072998046875, 0.2587890625, -0.27978515625, 0.356689453125, 0.2578125, -0.6396484375, 0.14892578125, 0.2802734375, -0.125732421875, -0.25, -0.2193603515625, -0.31689453125, -0.80126953125, 0.020263671875, -0.064453125, -0.10162353515625, -0.39501953125, 0.088134765625, -0.085205078125, 0.47021484375, -0.412109375, -1.404296875, -0.0728759765625, -0.058349609375, 0.17236328125, 0.1688232421875, -0.07708740234375, 0.3330078125, 0.0780029296875, -0.05712890625, 0.321044921875, 0.0091552734375, 0.2666015625, 0.43701171875, 0.0869140625, 0.1142578125, -0.1923828125, 0.28564453125, 0.1473388671875, -0.3564453125, -0.6455078125, -0.16455078125, -0.3359375, 0.050048828125, 0.2548828125, -0.260009765625, 0.83251953125, -0.0970458984375, 0.10791015625, -0.8759765625, 0.0697021484375, 0.0093994140625, 0.15673828125, -0.5810546875, 0.2578125, -0.607421875, 0.32666015625, -0.04913330078125, 0.263916015625, -0.0894775390625, -0.45751953125, -0.6201171875, 0.047149658203125, -0.45166015625, 0.475341796875, -0.218994140625, 0.98486328125, -0.2568359375, 0.098388671875, -0.130615234375, -0.07861328125, -0.42529296875, 0.021484375, 1.43359375, 0.032318115234375, 0.3330078125, -0.05029296875, 0.4365234375, -1.033203125, -0.153076171875, 0.521484375, 0.024658203125, -0.241943359375, 0.5390625, 0.1041259765625, 0.0908203125, 0.2109375, 0.007904052734375, 0.1826171875, -0.07861328125, 0.49072265625, -0.5595703125, 0.0869140625, -0.57568359375, 0.0643310546875, 0.0792236328125, -0.46630859375, -0.040283203125, -0.01666259765625, -0.1083984375, 0.117431640625, -0.7998046875, -0.22412109375, 0.33837890625, -0.0916748046875, 0.26611328125, 0.043701171875, -0.098388671875, -0.822265625, 0.0091552734375, 0.26904296875, 0.0299072265625, 0.1654052734375, -0.262451171875, -0.560546875, 0.52685546875, -0.420166015625, 0.0791015625, -0.56640625, 0.076171875, -0.433349609375, 0.09625244140625, 0.06396484375, 0.075439453125, 0.037139892578125, 0.2373046875, -2.2890625, 0.28662109375, -0.136962890625, 0.260009765625, 0.63037109375, 0.00146484375, -0.313232421875, -0.18994140625, 0.15283203125, -0.1279296875, 0.2265625, 0.01123046875, 0.15869140625, 0.5068359375, 0.33349609375, 0.365966796875, 0.051025390625, 0.3037109375, 0.59130859375, -0.279296875, 0.365478515625, 0.0648193359375, 0.008819580078125, -0.30810546875, 0.073974609375, 0.2490234375, 0.34130859375, -0.6982421875, 0.5810546875, 0.083984375, -0.1405029296875, 0.029541015625, -0.388671875, -0.0029296875, -0.01422119140625, 0.39794921875, -0.326171875, 0.04583740234375, -0.1788330078125, -0.181396484375, 0.415283203125, -0.00994873046875, -0.295166015625, -0.677734375, -0.345703125, 0.249267578125, 0.05859375, -0.213134765625, 0.31298828125, -0.0772705078125, 0.262451171875, -0.4453125, -0.302734375, 0.218017578125, -0.369873046875, -0.40478515625, -0.62890625, -0.60400390625, 0.016815185546875, 0.763671875, 0.673828125, -0.32666015625, 0.0023193359375, 0.65625, -0.599609375, 0.185791015625, -0.728515625, -0.7060546875, -0.48876953125, 0.16796875, 0.07073974609375, -0.329345703125, -0.2939453125, 0.1708984375, -0.10888671875, 0.16796875, 0.45361328125, 0.10479736328125, 0.1361083984375, 0.0166015625, -0.56787109375, 0.097412109375, -0.052001953125, -0.0413818359375, 0.76025390625, 0.1025390625, -0.339599609375, 0.45654296875, -0.5888671875, 0.441650390625, -0.2467041015625, 0.52099609375, -0.1484375, 0.06396484375, 0.1551513671875, -0.966796875, -0.1259765625, 0.052734375, 0.1805419921875, -0.3056640625, -0.41552734375, 0.0025634765625, 1.01953125, 0.371337890625, -0.62548828125, -0.2275390625, 0.6904296875, -0.125244140625, -0.18701171875, -0.034027099609375, -0.46240234375, 0.221435546875, -0.165771484375, -0.4619140625, 0.41552734375, 0.476318359375, -0.28369140625, -0.54443359375, 0.0699462890625, 0.08154296875, -0.090087890625, 0.1629638671875, -0.295654296875, -0.111083984375, 0.24853515625, 0.25927734375, 0.35302734375, 0.33154296875, 0.220703125, -0.087158203125, 0.17333984375, -0.01605224609375, 0.3564453125, 0.359375, 0.484375, -0.295166015625, -0.09130859375, 0.144775390625, -0.302490234375, -0.07177734375, -0.285888671875, 0.3251953125, 0.48291015625, -0.2529296875, 0.4853515625, -2.1171875, 0.498779296875, 0.43408203125, 0.157470703125, -0.44970703125, -0.218994140625, 0.118896484375, -0.1669921875, 0.038330078125, 0.2265625, 0.480712890625, -0.166748046875, 0.73046875, -0.3447265625, 0.0518798828125, 0.6787109375, 0.235595703125, 0.2705078125, -0.21337890625, -0.36474609375, 0.1788330078125, 0.4423828125, 0.1734619140625, -0.06005859375, -0.334228515625, 0.33544921875, 0.106201171875, 0.9638671875, 0.26171875, 0.333251953125, -0.301025390625, 0.361328125, 0.22900390625, -0.01123046875, 0.80029296875, -0.638671875, -0.273681640625, 0.404541015625, -0.63330078125, -0.0482177734375, 0.751953125, -0.44482421875, -0.2325439453125, -0.097412109375, -0.4384765625, -0.116455078125, 0.1820068359375, -0.99267578125, -0.049072265625, 0.07379150390625, -0.04296875, 0.6416015625, -0.3212890625, 0.419189453125, -0.275146484375, -0.2344970703125, -0.229248046875, -0.06396484375, 0.062744140625, -0.1400146484375, 0.16162109375, -0.0175323486328125, 0.10467529296875, 0.287353515625, -0.092041015625, -0.108642578125, -0.3818359375, -0.00960540771484375, 0.4482421875, 1.275390625, -0.1522216796875, -0.2333984375, -0.53759765625, 0.52783203125, 0.29248046875, 0.0548095703125, 0.29638671875, -0.279296875, -1.37890625, -0.0601806640625, 0.055908203125, -0.58984375, 0.416259765625, 0.54931640625, -0.0089111328125, 0.208740234375, -0.11065673828125, 0.07177734375, 0.13525390625, 0.03399658203125, -0.58837890625, 0.3359375, -0.11865234375, -0.490234375, 0.245361328125, -0.0963134765625, -0.331298828125, -0.1787109375, -0.136474609375, -0.100341796875, -0.47119140625, 0.04833984375, -0.0069580078125, -0.398193359375, -0.4931640625, -0.189453125, -0.4951171875, -0.1461181640625, 0.32421875, -0.074462890625, -0.253662109375, -0.193115234375, -0.19091796875, -0.1820068359375, 0.0908203125, 0.478271484375, 0.05078125, 0.1513671875, -0.6611328125, -0.318359375, -0.4892578125, -0.2421875, 0.0003662109375, 0.169921875, 0.2012939453125, 0.0413818359375, 0.42138671875, -0.0234527587890625, 0.1248779296875, -0.04034423828125, -0.2276611328125, 0.4345703125, 0.1346435546875, 0.58251953125, -0.122314453125, 0.0361328125, -0.7578125, -0.0670166015625, 0.00439453125, 0.17822265625, -0.25048828125, -0.0823974609375, -0.5361328125, 0.298095703125, -1.3984375, 0.1253662109375, -0.10992431640625, 1.251953125, -0.158447265625, 0.42724609375, 0.341552734375, 0.341796875, -0.2054443359375, -0.011474609375, 0.26171875, -0.033935546875, 0.4326171875, 0.5693359375, -0.54296875, -0.22607421875, -0.2001953125, -0.1285400390625, -0.03314208984375, 0.201171875, 0.43017578125, 0.05072021484375, -0.626953125, 0.232421875, -0.219482421875, -0.048126220703125, -0.69970703125, 0.8427734375, -0.0146484375, -0.0172119140625, 0.007080078125, -0.7275390625, 0.448974609375, 0.28955078125, 0.5283203125, -0.1240234375, -0.7353515625, 0.54248046875, 0.404296875, -0.1285400390625, 1.0947265625, -0.02197265625, -0.10498046875, -0.751953125, -0.14111328125, -0.205810546875, -0.0357666015625, 0.034912109375, 0.1549072265625, 0.5048828125, -0.18212890625, 0.1162109375, 0.038818359375, -0.5166015625, -0.277587890625, 0.205322265625, -0.44482421875, -0.29345703125, 0.1068115234375, -0.37548828125, -0.0523681640625, -0.1036376953125, 0.0245361328125, 0.8798828125, 0.054901123046875, -0.27734375, -0.150390625, 0.1788330078125, 0.0611572265625, 0.00787353515625, 0.25146484375, -0.1319580078125, 0.25537109375, 0.5244140625, -0.41455078125, 0.23193359375, -0.33837890625, -0.78955078125, 0.1539306640625, 0.270263671875, 0.18408203125, 0.05126953125, 0.076171875, 0.072998046875, 0.0946044921875, -0.216552734375, 0.3916015625, 0.69580078125]}, {"frame_number": 22, "vector": [-0.1907958984375, 0.27001953125, -0.0155029296875, -0.1719970703125, -0.62158203125, 0.31298828125, 0.0577392578125, -0.437255859375, -0.0679931640625, -0.040252685546875, -0.095947265625, 0.0478515625, 0.31103515625, 0.244873046875, 0.005859375, 0.26416015625, 0.11572265625, -0.14697265625, -0.06591796875, -0.46484375, -0.53857421875, -6.16015625, -0.28759765625, -0.72265625, 0.066162109375, -0.10546875, -0.1982421875, -0.876953125, 0.52490234375, -0.1903076171875, -0.9287109375, -0.6259765625, -0.6513671875, 0.0501708984375, -0.08038330078125, 0.0479736328125, 0.376708984375, -0.63525390625, -0.0479736328125, -0.4658203125, -0.04876708984375, 0.203369140625, 0.1796875, 0.0443115234375, 0.9931640625, -0.3203125, -0.385986328125, -0.1766357421875, -0.58740234375, -0.1611328125, -0.0196533203125, -0.2705078125, 0.0916748046875, 0.5009765625, 0.21923828125, 0.012847900390625, 0.40283203125, -0.246826171875, 0.26123046875, 0.059326171875, -0.07568359375, 0.1109619140625, 0.24072265625, 0.18017578125, -0.308349609375, -0.3359375, -0.49365234375, 0.0185546875, 0.23779296875, 0.0604248046875, 0.9130859375, -0.0260009765625, -0.0146484375, -0.039306640625, -0.41064453125, 0.83251953125, -0.33349609375, -0.44775390625, 0.1962890625, -0.7177734375, 1.16796875, 0.32666015625, 0.5576171875, 0.472412109375, -0.275390625, 0.08026123046875, 0.1864013671875, -0.0606689453125, 0.2490234375, -0.638671875, -0.08819580078125, -0.1112060546875, 0.025390625, -0.1455078125, 0.3232421875, 0.37255859375, 0.0751953125, 0.46826171875, 0.03692626953125, -0.67333984375, 0.72900390625, 0.285400390625, -0.4072265625, -0.6611328125, 0.03466796875, -0.074462890625, 0.02203369140625, -0.133544921875, -0.156005859375, -0.051025390625, -0.0955810546875, 0.1171875, 0.127197265625, -0.4326171875, 0.134765625, -0.53515625, -0.1689453125, -0.0887451171875, 0.3310546875, 0.5419921875, 0.069091796875, -0.302734375, -0.033203125, -0.6943359375, -0.054443359375, 0.363037109375, -0.1334228515625, 0.33984375, 0.1884765625, 0.271728515625, -0.01220703125, -0.4580078125, -0.053955078125, -0.10894775390625, 0.53662109375, -0.0675048828125, 0.0880126953125, 0.7646484375, 0.5244140625, 0.0400390625, 0.0198974609375, 0.342529296875, 0.1376953125, -0.1202392578125, -0.22802734375, 0.096435546875, -0.0272216796875, -0.388671875, 0.2109375, 1.078125, -0.1617431640625, -1.5166015625, 0.1395263671875, -0.42333984375, 0.426513671875, -0.274658203125, -0.18212890625, 0.99365234375, 0.3544921875, -0.409423828125, -0.2958984375, 0.224609375, 0.39599609375, 0.422607421875, -0.2734375, -0.270263671875, 0.013427734375, 0.64599609375, 0.0079345703125, 0.0421142578125, 0.163818359375, 0.433349609375, -0.175537109375, 0.5625, -0.285400390625, 1.0107421875, -1.107421875, 0.1282958984375, -0.013916015625, -0.071044921875, 0.0545654296875, 0.1259765625, -0.20458984375, 0.08941650390625, 0.0052490234375, 0.026611328125, -0.0322265625, 0.1961669921875, -0.0809326171875, 0.07391357421875, 0.2003173828125, 0.1806640625, 0.4873046875, -0.37255859375, 0.556640625, -0.2008056640625, 0.44873046875, -0.153564453125, -0.2783203125, 1.2041015625, 0.492431640625, 0.0, -0.171142578125, 0.0677490234375, -0.044677734375, 0.0604248046875, 0.62255859375, 0.015289306640625, -0.62109375, -0.16796875, 0.1590576171875, -0.02880859375, 0.257568359375, -0.86474609375, -0.4990234375, -0.5537109375, 0.48388671875, 0.99951171875, -0.02374267578125, -0.448974609375, -0.444091796875, -0.52978515625, -0.1446533203125, -0.56005859375, -0.15185546875, 0.2099609375, -0.1434326171875, -0.2451171875, 0.7109375, 0.0567626953125, -0.037353515625, 0.278564453125, 0.0595703125, 0.314208984375, -0.42529296875, -0.5341796875, 0.705078125, -0.044921875, -1.044921875, -0.075439453125, 0.66796875, 0.2373046875, -0.07550048828125, 0.203369140625, -0.1492919921875, 1.015625, -0.2216796875, -0.0263671875, -0.25537109375, -0.109130859375, 0.000244140625, -0.0340576171875, 0.29638671875, -0.070068359375, -0.6337890625, -0.035888671875, -0.10400390625, -0.05419921875, 0.38427734375, 0.4638671875, 0.0010986328125, -0.195068359375, -0.0595703125, -0.0233154296875, -0.259765625, -0.051025390625, -0.82275390625, -0.291015625, 0.58349609375, -0.394287109375, -0.017364501953125, 0.4521484375, 0.353271484375, 0.28369140625, -0.2291259765625, -0.66943359375, -0.197021484375, 0.158203125, -0.346923828125, -0.3720703125, -0.1815185546875, 0.671875, -0.37939453125, 0.166259765625, 0.4189453125, 0.289306640625, -0.153076171875, 0.52880859375, -0.556640625, 1.1171875, -0.031982421875, 0.07501220703125, 0.03173828125, -0.467041015625, 0.0552978515625, 0.1121826171875, 0.103515625, 0.36572265625, 0.578125, -0.29052734375, 0.004840850830078125, 0.740234375, -0.06158447265625, -0.009521484375, -0.308349609375, 0.431884765625, 0.325927734375, -0.654296875, 0.120849609375, -0.0811767578125, 0.3486328125, -0.373291015625, -0.0068359375, -0.265625, -0.3291015625, 0.340576171875, 0.0499267578125, 0.03057861328125, -0.24951171875, 0.11181640625, -0.8974609375, 0.16064453125, 0.08349609375, -0.4208984375, -0.9052734375, 0.591796875, -0.4384765625, 0.27685546875, -0.5869140625, -0.057373046875, -0.509765625, -0.0843505859375, -0.105712890625, -0.4482421875, 0.166015625, 0.281982421875, 0.1649169921875, -0.59765625, 0.006103515625, -0.0205078125, -0.77587890625, 0.113037109375, -0.3173828125, -0.677734375, -0.595703125, 0.158935546875, -0.412109375, 0.373291015625, 0.177978515625, 0.0191650390625, -0.299560546875, -0.32080078125, 0.298095703125, 0.190185546875, 0.37841796875, 0.30615234375, 0.247802734375, -0.1693115234375, -0.33984375, -0.1959228515625, -0.212646484375, -0.05712890625, 0.5400390625, 0.271240234375, -0.6044921875, -0.499267578125, -0.0048828125, 0.2386474609375, 0.03900146484375, -0.153564453125, -0.0130615234375, 0.216796875, -0.071533203125, 0.3232421875, 0.306884765625, -0.0848388671875, 0.43017578125, -0.22265625, 0.05517578125, -0.5576171875, -0.0391845703125, -0.30322265625, 0.1380615234375, 0.7275390625, 0.181640625, -0.0418701171875, -0.03204345703125, -0.167236328125, -0.263671875, 0.4990234375, -0.6279296875, -0.10107421875, 0.2734375, 0.25537109375, 0.9775390625, 0.64794921875, -0.144287109375, -0.69775390625, -0.080810546875, -0.33837890625, 0.18798828125, 0.2293701171875, -0.118408203125, -0.39794921875, -0.14404296875, 0.028564453125, -0.36083984375, 0.05322265625, 0.3662109375, -0.111083984375, 0.2529296875, -0.89404296875, -0.70556640625, -0.28857421875, -0.331298828125, 0.493408203125, 0.7431640625, -0.24365234375, 1.205078125, -0.26708984375, 0.23779296875, -0.0794677734375, -0.19189453125, -0.0631103515625, -0.0107421875, -0.50927734375, -1.318359375, -0.043701171875, 0.1539306640625, 0.230224609375, 0.0853271484375, -6.171875, -0.3681640625, -0.07958984375, -0.77490234375, -0.2548828125, -0.105712890625, -1.1796875, -0.68798828125, -0.250732421875, -0.0107421875, 0.6787109375, 0.64404296875, -0.17138671875, 0.00555419921875, 0.2001953125, 0.2353515625, -0.0078125, -0.129150390625, 0.18115234375, 0.022705078125, 0.2117919921875, -0.0931396484375, 0.20068359375, 0.463623046875, -0.26904296875, -0.1683349609375, 0.304443359375, -0.63818359375, -0.66552734375, -0.0908203125, 0.0400390625, -0.3515625, -0.294677734375, -0.451171875, -0.030029296875, 0.60400390625, 0.7587890625, -0.23388671875, -0.46533203125, 0.2335205078125, -0.99560546875, 0.31396484375, -0.0347900390625, -0.072998046875, 0.1937255859375, 0.332275390625, -0.0977783203125, -0.71923828125, -0.77294921875, 0.140625, 0.13623046875, -1.0625, 0.006591796875, 0.29736328125, 0.0032958984375, 0.278564453125, -0.10382080078125, 0.1165771484375, 0.172607421875, 0.229248046875, 0.57177734375, 0.318603515625, 0.425048828125, -0.381591796875, 0.04248046875, 0.047882080078125, 0.0228271484375, 0.1142578125, 0.01513671875, -0.0491943359375, -0.00909423828125, 0.64892578125, -0.23828125, -0.3515625, -0.06787109375, -0.066650390625, 0.021240234375, -0.10595703125, 0.12255859375, 0.2098388671875, 0.049072265625, 0.1279296875, 0.282470703125, -0.2275390625, -0.242431640625, -0.114990234375, -0.380126953125, -0.6494140625, -0.0712890625, 0.6572265625, 0.14794921875, -0.2305908203125, 0.053558349609375, 0.2421875, 1.26171875, 0.1708984375, 0.224609375, 0.0936279296875, 0.10595703125, -0.38037109375, -0.474853515625, 0.025634765625, -0.041748046875, 0.337890625, 0.241455078125, 0.07470703125, 0.0028076171875, 0.4404296875, 0.03564453125, 0.6513671875, 0.1112060546875, 0.416015625, 0.52685546875, -0.1595458984375, 0.513671875, -1.5283203125, -0.3330078125, -0.259033203125, 0.1181640625, -0.048583984375, -0.261962890625, 0.7880859375, 0.31396484375, -0.0399169921875, 0.06591796875, 0.302001953125, -0.1082763671875, -0.24560546875, -0.99853515625, -0.438720703125, -0.29248046875, -0.380126953125, -0.8701171875, 0.369140625, 0.260009765625, 0.455078125, -0.37646484375, -0.145263671875, 0.1104736328125, -0.0843505859375, 0.3525390625, -0.251953125, 0.1224365234375, 0.2010498046875, 0.463623046875, -0.43798828125, 0.2105712890625, -0.31884765625, -0.467529296875, -0.39013671875, 0.1376953125, -0.200439453125, -0.12646484375, 0.0506591796875, -0.0809326171875, 0.02783203125, 0.484130859375, -0.2470703125, 0.208251953125, 0.3896484375, -0.0286865234375, -0.2177734375, 0.0194091796875, 0.07366943359375, 0.0994873046875, -0.1494140625, -0.25732421875, -1.4873046875, -0.4775390625, 0.175048828125, -0.0123291015625, -0.3466796875, -0.40673828125, -0.23193359375, -0.079833984375, -0.0159912109375, -0.219970703125, -0.202392578125, -0.88330078125, 0.17041015625, 0.326904296875, -0.1971435546875, -0.189453125, -0.1767578125, -0.5478515625, 0.25927734375, 0.292724609375, 0.0228271484375, -0.180419921875, 0.0223388671875, -0.09521484375, 0.04510498046875, 0.06610107421875, -0.278564453125, 0.020721435546875, 0.03680419921875, 0.3818359375, 0.230712890625, -0.0946044921875, 0.221923828125, -0.0245361328125, -0.3916015625, 0.701171875, 0.091552734375, -0.18310546875, -0.29833984375, 1.314453125, 0.09033203125, -0.31298828125, 0.099609375, -0.3583984375, 0.353759765625, -0.206298828125, -0.131591796875, 0.3525390625, 0.1522216796875, 0.022705078125, 0.234130859375, 0.09832763671875, 0.293212890625, -0.306884765625, 0.3310546875, 0.2509765625, -0.63037109375, 0.163330078125, 0.299560546875, -0.1177978515625, -0.23681640625, -0.19287109375, -0.34033203125, -0.82568359375, -0.00347900390625, -0.06414794921875, -0.07635498046875, -0.39990234375, 0.04931640625, -0.087890625, 0.41650390625, -0.422119140625, -1.392578125, -0.04541015625, -0.1424560546875, 0.16748046875, 0.1937255859375, -0.0789794921875, 0.3330078125, 0.1317138671875, -0.051513671875, 0.30126953125, 0.01873779296875, 0.234130859375, 0.44287109375, 0.062744140625, 0.11529541015625, -0.108154296875, 0.29931640625, 0.1470947265625, -0.3759765625, -0.5859375, -0.14599609375, -0.347900390625, 0.020751953125, 0.271484375, -0.310302734375, 0.9404296875, -0.080078125, 0.136474609375, -0.84912109375, 0.0086669921875, -0.05255126953125, 0.142333984375, -0.55078125, 0.2958984375, -0.595703125, 0.38134765625, -0.05523681640625, 0.27392578125, -0.1121826171875, -0.50244140625, -0.6748046875, 0.09442138671875, -0.45556640625, 0.47802734375, -0.251953125, 1.0126953125, -0.32568359375, 0.04876708984375, -0.134521484375, -0.0758056640625, -0.41552734375, -0.0211181640625, 1.408203125, 0.0687255859375, 0.314453125, -0.033935546875, 0.3955078125, -1.091796875, -0.2022705078125, 0.5, 0.03607177734375, -0.218017578125, 0.5546875, 0.12109375, 0.090087890625, 0.2001953125, 0.0545654296875, 0.16162109375, -0.041748046875, 0.51513671875, -0.489990234375, 0.056640625, -0.6064453125, 0.0107421875, 0.11865234375, -0.5166015625, -0.05462646484375, -0.04827880859375, -0.122802734375, 0.131103515625, -0.8046875, -0.2301025390625, 0.318115234375, -0.08349609375, 0.274169921875, -0.013427734375, -0.097900390625, -0.7763671875, 0.060791015625, 0.260986328125, 0.002685546875, 0.12744140625, -0.25634765625, -0.59912109375, 0.53662109375, -0.455322265625, 0.033935546875, -0.6162109375, 0.07135009765625, -0.43408203125, 0.09716796875, 0.0234375, 0.046630859375, 0.0345458984375, 0.2203369140625, -2.25390625, 0.2900390625, -0.211181640625, 0.259765625, 0.669921875, -0.0343017578125, -0.357666015625, -0.19287109375, 0.21240234375, -0.1341552734375, 0.231689453125, 0.08642578125, 0.142822265625, 0.5439453125, 0.383544921875, 0.390380859375, -0.02734375, 0.26171875, 0.62890625, -0.31591796875, 0.26953125, 0.03753662109375, 0.01702880859375, -0.295654296875, 0.057861328125, 0.25732421875, 0.3681640625, -0.77294921875, 0.5966796875, 0.0711669921875, -0.1875, 0.0440673828125, -0.46630859375, 0.0194091796875, -0.0509033203125, 0.34375, -0.278076171875, -0.004119873046875, -0.169921875, -0.1904296875, 0.47216796875, -0.028411865234375, -0.341064453125, -0.625, -0.347900390625, 0.259765625, 0.077392578125, -0.212158203125, 0.278076171875, -0.072265625, 0.26806640625, -0.4228515625, -0.27734375, 0.2215576171875, -0.3740234375, -0.471435546875, -0.666015625, -0.59765625, 0.01629638671875, 0.7568359375, 0.66259765625, -0.2509765625, 0.03692626953125, 0.7080078125, -0.5654296875, 0.26611328125, -0.6865234375, -0.6923828125, -0.533203125, 0.152099609375, 0.08319091796875, -0.32275390625, -0.30029296875, 0.2103271484375, -0.1318359375, 0.12353515625, 0.436279296875, 0.02001953125, 0.09912109375, 0.010345458984375, -0.564453125, 0.096435546875, 0.00762939453125, -0.0621337890625, 0.80859375, 0.086181640625, -0.3232421875, 0.454345703125, -0.6318359375, 0.44580078125, -0.29443359375, 0.572265625, -0.1322021484375, 0.0830078125, 0.147705078125, -0.9814453125, -0.137451171875, 0.018798828125, 0.1939697265625, -0.24658203125, -0.44384765625, 0.01123046875, 0.978515625, 0.3876953125, -0.5703125, -0.2216796875, 0.7119140625, -0.1103515625, -0.17431640625, -0.06292724609375, -0.43603515625, 0.2425537109375, -0.1671142578125, -0.4501953125, 0.442626953125, 0.443603515625, -0.271484375, -0.54443359375, 0.0694580078125, 0.125244140625, -0.12213134765625, 0.164306640625, -0.25, -0.1019287109375, 0.15185546875, 0.308349609375, 0.34423828125, 0.38671875, 0.19873046875, -0.08984375, 0.141845703125, -0.0562744140625, 0.354248046875, 0.326171875, 0.4892578125, -0.297607421875, -0.11029052734375, 0.159423828125, -0.29052734375, -0.02392578125, -0.25244140625, 0.325927734375, 0.43701171875, -0.25830078125, 0.529296875, -2.10546875, 0.52001953125, 0.41015625, 0.100830078125, -0.400634765625, -0.16650390625, 0.09423828125, -0.12646484375, 0.03924560546875, 0.298828125, 0.3994140625, -0.179931640625, 0.77685546875, -0.349609375, -0.0341796875, 0.63671875, 0.2271728515625, 0.2900390625, -0.2216796875, -0.363525390625, 0.203857421875, 0.49072265625, 0.1761474609375, -0.0693359375, -0.2919921875, 0.296142578125, 0.11761474609375, 0.92138671875, 0.28369140625, 0.396728515625, -0.263671875, 0.376953125, 0.210693359375, -0.022064208984375, 0.650390625, -0.662109375, -0.302734375, 0.4697265625, -0.6201171875, -0.0328369140625, 0.787109375, -0.44970703125, -0.2318115234375, -0.09326171875, -0.44482421875, -0.181396484375, 0.19873046875, -0.94189453125, -0.0916748046875, 0.094482421875, -0.0316162109375, 0.65966796875, -0.3154296875, 0.40380859375, -0.2978515625, -0.189453125, -0.2271728515625, -0.057373046875, 0.052734375, -0.0968017578125, 0.1165771484375, 7.62939453125e-05, 0.0469970703125, 0.29736328125, -0.084228515625, -0.141357421875, -0.344482421875, 0.0034637451171875, 0.40869140625, 1.244140625, -0.1551513671875, -0.21533203125, -0.55126953125, 0.49951171875, 0.29931640625, 0.0550537109375, 0.277587890625, -0.27783203125, -1.30859375, -0.0535888671875, 0.0345458984375, -0.59521484375, 0.43603515625, 0.5185546875, 0.0369873046875, 0.2275390625, -0.130126953125, 0.127197265625, 0.177734375, 0.04559326171875, -0.5380859375, 0.36328125, -0.141845703125, -0.474365234375, 0.191650390625, -0.110595703125, -0.337890625, -0.211669921875, -0.16162109375, -0.0986328125, -0.44384765625, 0.04150390625, 0.0135498046875, -0.38818359375, -0.53759765625, -0.16943359375, -0.5419921875, -0.12042236328125, 0.28369140625, -0.016357421875, -0.2451171875, -0.19482421875, -0.171630859375, -0.1519775390625, 0.15478515625, 0.54052734375, 0.06689453125, 0.173583984375, -0.6767578125, -0.321044921875, -0.481689453125, -0.250244140625, 0.015380859375, 0.2188720703125, 0.185791015625, 0.044677734375, 0.416015625, -0.029022216796875, 0.182373046875, -0.03173828125, -0.2294921875, 0.365478515625, 0.187744140625, 0.57373046875, -0.136962890625, 0.05224609375, -0.78271484375, -0.0467529296875, 0.04443359375, 0.185302734375, -0.282470703125, -0.06280517578125, -0.55419921875, 0.28759765625, -1.4208984375, 0.17626953125, -0.09637451171875, 1.2294921875, -0.1026611328125, 0.413330078125, 0.35546875, 0.441650390625, -0.1756591796875, -0.0255126953125, 0.2802734375, -0.0181884765625, 0.3701171875, 0.5283203125, -0.5693359375, -0.225830078125, -0.2100830078125, -0.1256103515625, -0.04925537109375, 0.1539306640625, 0.4375, 0.06976318359375, -0.666015625, 0.2069091796875, -0.244384765625, -0.05712890625, -0.712890625, 0.84716796875, -0.01220703125, -0.02685546875, 0.0244140625, -0.759765625, 0.455322265625, 0.2423095703125, 0.537109375, -0.119384765625, -0.70849609375, 0.58740234375, 0.39306640625, -0.0888671875, 1.1015625, -0.00732421875, -0.1116943359375, -0.7255859375, -0.16162109375, -0.21142578125, 0.0050048828125, -0.0498046875, 0.140869140625, 0.45849609375, -0.16357421875, 0.09967041015625, 0.0401611328125, -0.52734375, -0.307861328125, 0.20361328125, -0.46630859375, -0.2880859375, 0.10784912109375, -0.344970703125, -0.09161376953125, -0.11737060546875, -0.03387451171875, 0.85986328125, 0.08282470703125, -0.2266845703125, -0.158935546875, 0.1915283203125, 0.103759765625, -0.00335693359375, 0.21728515625, -0.13134765625, 0.26953125, 0.55224609375, -0.38916015625, 0.1484375, -0.345458984375, -0.81884765625, 0.1156005859375, 0.2431640625, 0.203125, 0.0513916015625, 0.071533203125, 0.11962890625, 0.1065673828125, -0.193359375, 0.385498046875, 0.69189453125]}, {"frame_number": 23, "vector": [-0.22119140625, 0.257568359375, -0.006591796875, -0.1729736328125, -0.5888671875, 0.33203125, 0.0892333984375, -0.448974609375, -0.065673828125, -0.07684326171875, -0.0869140625, 0.03857421875, 0.3203125, 0.27197265625, 0.0303955078125, 0.28662109375, 0.10809326171875, -0.158203125, -0.06671142578125, -0.48876953125, -0.59326171875, -6.1015625, -0.3359375, -0.70849609375, 0.04730224609375, -0.080810546875, -0.208984375, -0.86865234375, 0.49755859375, -0.187255859375, -0.9287109375, -0.61083984375, -0.6591796875, 0.0504150390625, -0.0999755859375, 0.02545166015625, 0.35595703125, -0.6708984375, -0.06121826171875, -0.4443359375, -0.062744140625, 0.1968994140625, 0.169189453125, 0.0440673828125, 1.0048828125, -0.38330078125, -0.38134765625, -0.1646728515625, -0.5791015625, -0.140625, -0.006195068359375, -0.282470703125, 0.1221923828125, 0.489990234375, 0.228271484375, -0.003265380859375, 0.3994140625, -0.27490234375, 0.278076171875, 0.05975341796875, -0.04541015625, 0.126708984375, 0.259521484375, 0.15966796875, -0.29443359375, -0.3095703125, -0.4931640625, 0.013916015625, 0.25244140625, 0.1065673828125, 0.92529296875, -0.0281982421875, 0.011474609375, -0.03759765625, -0.40673828125, 0.82373046875, -0.33447265625, -0.4296875, 0.21484375, -0.6708984375, 1.1484375, 0.32080078125, 0.57421875, 0.46533203125, -0.267578125, 0.078125, 0.2137451171875, -0.020721435546875, 0.1962890625, -0.650390625, -0.09918212890625, -0.10198974609375, 0.01513671875, -0.106201171875, 0.293701171875, 0.36572265625, 0.080078125, 0.486328125, 0.0562744140625, -0.654296875, 0.7451171875, 0.31787109375, -0.4169921875, -0.6767578125, 0.0089111328125, -0.0721435546875, 0.052337646484375, -0.1351318359375, -0.160400390625, -0.036376953125, -0.117431640625, 0.123291015625, 0.0941162109375, -0.40234375, 0.1229248046875, -0.5244140625, -0.117919921875, -0.0966796875, 0.31689453125, 0.53515625, 0.07470703125, -0.290771484375, -0.04888916015625, -0.63671875, -0.0225830078125, 0.345947265625, -0.164794921875, 0.383056640625, 0.12939453125, 0.28369140625, -0.00244140625, -0.44140625, -0.06201171875, -0.061187744140625, 0.5029296875, -0.053466796875, 0.1092529296875, 0.77734375, 0.5498046875, 0.0576171875, -0.0020751953125, 0.325439453125, 0.1571044921875, -0.13427734375, -0.25634765625, 0.110107421875, -0.0313720703125, -0.407470703125, 0.218017578125, 1.06640625, -0.1527099609375, -1.5703125, 0.137451171875, -0.47607421875, 0.412841796875, -0.280029296875, -0.169921875, 1.0, 0.343017578125, -0.439697265625, -0.294921875, 0.21923828125, 0.39453125, 0.381591796875, -0.29833984375, -0.236572265625, -0.0003662109375, 0.65771484375, 0.0345458984375, -0.0126953125, 0.1492919921875, 0.4521484375, -0.14794921875, 0.5830078125, -0.2890625, 0.984375, -1.064453125, 0.13720703125, 0.0084228515625, -0.068359375, 0.0601806640625, 0.1173095703125, -0.1904296875, 0.03826904296875, 0.007568359375, 0.016845703125, -0.040771484375, 0.178466796875, -0.08056640625, 0.0980224609375, 0.1868896484375, 0.160888671875, 0.483642578125, -0.39794921875, 0.5361328125, -0.237548828125, 0.46240234375, -0.1875, -0.260009765625, 1.208984375, 0.5068359375, -0.03466796875, -0.1650390625, 0.055908203125, -0.00201416015625, 0.090576171875, 0.6416015625, 0.00958251953125, -0.62548828125, -0.1751708984375, 0.1773681640625, -0.06201171875, 0.227294921875, -0.853515625, -0.53466796875, -0.54736328125, 0.5166015625, 1.0224609375, -0.028350830078125, -0.450927734375, -0.44287109375, -0.50732421875, -0.118408203125, -0.564453125, -0.142578125, 0.1845703125, -0.11651611328125, -0.21484375, 0.69384765625, 0.05572509765625, -0.05718994140625, 0.24267578125, 0.045654296875, 0.333984375, -0.412109375, -0.54052734375, 0.71484375, -0.0572509765625, -0.97705078125, -0.10205078125, 0.72412109375, 0.272705078125, -0.061676025390625, 0.20654296875, -0.167724609375, 1.01171875, -0.212890625, -0.0115966796875, -0.2498779296875, -0.109619140625, 0.0511474609375, -0.047119140625, 0.31005859375, -0.06640625, -0.67919921875, -0.0615234375, -0.1083984375, -0.02490234375, 0.44091796875, 0.458740234375, -0.04107666015625, -0.2117919921875, -0.0384521484375, -0.0343017578125, -0.2529296875, -0.0335693359375, -0.8095703125, -0.318359375, 0.5361328125, -0.34912109375, 0.0113525390625, 0.4619140625, 0.337646484375, 0.272705078125, -0.215087890625, -0.68017578125, -0.1632080078125, 0.149169921875, -0.34716796875, -0.3994140625, -0.253173828125, 0.677734375, -0.375, 0.150390625, 0.39990234375, 0.267578125, -0.146728515625, 0.5380859375, -0.546875, 1.1455078125, -0.03009033203125, 0.11260986328125, 0.0218505859375, -0.4111328125, 0.0496826171875, 0.0872802734375, 0.116943359375, 0.35400390625, 0.6103515625, -0.267333984375, 0.01515960693359375, 0.73193359375, -0.0775146484375, -0.01806640625, -0.324462890625, 0.403076171875, 0.318359375, -0.65625, 0.1168212890625, -0.087890625, 0.3486328125, -0.357421875, 0.0185546875, -0.264892578125, -0.33837890625, 0.34912109375, 0.01568603515625, 0.0533447265625, -0.235107421875, 0.105224609375, -0.95458984375, 0.1712646484375, 0.11138916015625, -0.43701171875, -0.87255859375, 0.54443359375, -0.41552734375, 0.2493896484375, -0.5625, -0.057647705078125, -0.541015625, -0.0762939453125, -0.12255859375, -0.487548828125, 0.1470947265625, 0.28759765625, 0.1507568359375, -0.65771484375, 0.041290283203125, -0.0255126953125, -0.78173828125, 0.138916015625, -0.304931640625, -0.6884765625, -0.611328125, 0.185791015625, -0.42578125, 0.39599609375, 0.1865234375, -0.0098876953125, -0.3125, -0.303466796875, 0.310791015625, 0.1826171875, 0.396728515625, 0.32568359375, 0.2408447265625, -0.178955078125, -0.32958984375, -0.24609375, -0.20703125, -0.0416259765625, 0.5224609375, 0.27880859375, -0.60888671875, -0.43994140625, 0.0111083984375, 0.2442626953125, 0.013427734375, -0.1282958984375, -0.00262451171875, 0.22412109375, -0.100341796875, 0.339111328125, 0.3349609375, -0.09765625, 0.421875, -0.2431640625, 0.072265625, -0.60986328125, -0.064697265625, -0.296630859375, 0.1302490234375, 0.73828125, 0.1546630859375, -0.0389404296875, -0.0321044921875, -0.1103515625, -0.2484130859375, 0.5009765625, -0.625, -0.123779296875, 0.2705078125, 0.2529296875, 0.99072265625, 0.625, -0.139892578125, -0.7060546875, -0.083984375, -0.30419921875, 0.1982421875, 0.24853515625, -0.1220703125, -0.39599609375, -0.14501953125, 0.0247802734375, -0.351806640625, 0.02996826171875, 0.377685546875, -0.1025390625, 0.23974609375, -0.90771484375, -0.6845703125, -0.2646484375, -0.350341796875, 0.5146484375, 0.755859375, -0.271484375, 1.158203125, -0.2646484375, 0.1864013671875, -0.08526611328125, -0.231689453125, -0.0672607421875, 0.0225830078125, -0.52294921875, -1.3642578125, -0.045166015625, 0.1629638671875, 0.2154541015625, 0.04180908203125, -6.1171875, -0.3515625, -0.055419921875, -0.78076171875, -0.26171875, -0.1064453125, -1.208984375, -0.69140625, -0.252197265625, -0.021484375, 0.650390625, 0.615234375, -0.221923828125, 0.00970458984375, 0.2279052734375, 0.18896484375, -0.0277099609375, -0.167724609375, 0.187744140625, -0.031494140625, 0.2069091796875, -0.0819091796875, 0.1978759765625, 0.4521484375, -0.265625, -0.15087890625, 0.330322265625, -0.595703125, -0.64794921875, -0.114013671875, 0.052734375, -0.3369140625, -0.2939453125, -0.47412109375, -0.0576171875, 0.6259765625, 0.7626953125, -0.265380859375, -0.4306640625, 0.2349853515625, -0.97607421875, 0.3154296875, -0.0255126953125, -0.0865478515625, 0.19677734375, 0.352294921875, -0.1148681640625, -0.7607421875, -0.7841796875, 0.12890625, 0.1103515625, -1.068359375, -0.04443359375, 0.31201171875, -0.01837158203125, 0.272216796875, -0.10333251953125, 0.07879638671875, 0.205078125, 0.240234375, 0.53515625, 0.311767578125, 0.379638671875, -0.427978515625, 0.0380859375, 0.05487060546875, 0.0263671875, 0.11181640625, 0.0196533203125, -0.0670166015625, -0.022857666015625, 0.693359375, -0.238525390625, -0.369140625, -0.05517578125, -0.053466796875, 0.068603515625, -0.080322265625, 0.1025390625, 0.2203369140625, 0.0601806640625, 0.11016845703125, 0.306640625, -0.251708984375, -0.21044921875, -0.08349609375, -0.407470703125, -0.658203125, -0.07958984375, 0.677734375, 0.2265625, -0.2003173828125, 0.054840087890625, 0.255126953125, 1.2216796875, 0.16064453125, 0.2164306640625, 0.0765380859375, 0.1444091796875, -0.402099609375, -0.44921875, 0.032470703125, -0.0263671875, 0.372802734375, 0.224609375, 0.051513671875, -0.00579833984375, 0.3955078125, 0.0269775390625, 0.6689453125, 0.0821533203125, 0.3828125, 0.5400390625, -0.177978515625, 0.47900390625, -1.50390625, -0.341796875, -0.2481689453125, 0.136962890625, -0.022705078125, -0.2763671875, 0.73681640625, 0.309326171875, -0.05694580078125, 0.039306640625, 0.3232421875, -0.0921630859375, -0.262939453125, -0.9697265625, -0.47607421875, -0.312255859375, -0.376953125, -0.84912109375, 0.358642578125, 0.273681640625, 0.489990234375, -0.37353515625, -0.14697265625, 0.09576416015625, -0.0606689453125, 0.3466796875, -0.248046875, 0.1407470703125, 0.197265625, 0.448486328125, -0.4306640625, 0.2222900390625, -0.33447265625, -0.484130859375, -0.388427734375, 0.1361083984375, -0.191162109375, -0.1248779296875, 0.016357421875, -0.083740234375, 0.0263671875, 0.482421875, -0.24365234375, 0.236328125, 0.3818359375, 0.02691650390625, -0.1461181640625, 0.019775390625, 0.1025390625, 0.1048583984375, -0.140380859375, -0.2445068359375, -1.5068359375, -0.44775390625, 0.21923828125, -0.0330810546875, -0.3359375, -0.42041015625, -0.183837890625, -0.07763671875, -0.040008544921875, -0.2255859375, -0.216064453125, -0.875, 0.1583251953125, 0.3447265625, -0.2115478515625, -0.216796875, -0.136962890625, -0.55078125, 0.257568359375, 0.28271484375, 0.0496826171875, -0.2049560546875, 0.05419921875, -0.1162109375, 0.05059814453125, 0.06683349609375, -0.30078125, 0.0071258544921875, 0.01531982421875, 0.33056640625, 0.197509765625, -0.07373046875, 0.21337890625, -0.0215606689453125, -0.39990234375, 0.6435546875, 0.1099853515625, -0.1845703125, -0.299560546875, 1.306640625, 0.099609375, -0.3017578125, 0.08837890625, -0.38037109375, 0.348388671875, -0.22412109375, -0.153564453125, 0.31982421875, 0.1546630859375, 0.03314208984375, 0.287353515625, 0.10491943359375, 0.28515625, -0.3095703125, 0.326171875, 0.266357421875, -0.64990234375, 0.1435546875, 0.291748046875, -0.1112060546875, -0.24169921875, -0.208251953125, -0.3427734375, -0.822265625, 0.00390625, -0.038330078125, -0.08563232421875, -0.42529296875, 0.066162109375, -0.069580078125, 0.47412109375, -0.435546875, -1.33203125, -0.0516357421875, -0.0977783203125, 0.1787109375, 0.18505859375, -0.0665283203125, 0.373046875, 0.1019287109375, -0.0625, 0.288818359375, 0.00079345703125, 0.271240234375, 0.439208984375, 0.05322265625, 0.1103515625, -0.13623046875, 0.2919921875, 0.16796875, -0.389892578125, -0.6416015625, -0.1671142578125, -0.33740234375, 0.036376953125, 0.240478515625, -0.2978515625, 0.8623046875, -0.08154296875, 0.1234130859375, -0.8583984375, 0.01806640625, -0.0361328125, 0.14453125, -0.57080078125, 0.265869140625, -0.603515625, 0.3720703125, -0.0556640625, 0.29150390625, -0.0809326171875, -0.44189453125, -0.65087890625, 0.085693359375, -0.483154296875, 0.490478515625, -0.231689453125, 0.99755859375, -0.2841796875, 0.0799560546875, -0.154052734375, -0.071044921875, -0.41259765625, 0.014892578125, 1.396484375, 0.054962158203125, 0.300048828125, -0.0341796875, 0.42529296875, -1.1201171875, -0.17333984375, 0.5009765625, 0.05828857421875, -0.247314453125, 0.5205078125, 0.1405029296875, 0.10986328125, 0.185546875, 0.04669189453125, 0.16650390625, -0.064208984375, 0.52099609375, -0.5390625, 0.0947265625, -0.59423828125, 0.0548095703125, 0.0911865234375, -0.5048828125, -0.0540771484375, -0.009033203125, -0.1112060546875, 0.114990234375, -0.8037109375, -0.247314453125, 0.323974609375, -0.085693359375, 0.26220703125, -0.0001220703125, -0.1009521484375, -0.8046875, 0.0574951171875, 0.25341796875, 0.0413818359375, 0.1346435546875, -0.283447265625, -0.5869140625, 0.55029296875, -0.436767578125, 0.0517578125, -0.60302734375, 0.063720703125, -0.452880859375, 0.1025390625, 0.0263671875, 0.092041015625, 0.033203125, 0.2412109375, -2.263671875, 0.30859375, -0.16064453125, 0.27880859375, 0.64990234375, -0.024169921875, -0.34912109375, -0.1884765625, 0.18798828125, -0.1337890625, 0.25146484375, 0.044677734375, 0.166015625, 0.5517578125, 0.3671875, 0.380126953125, 0.0335693359375, 0.2958984375, 0.6103515625, -0.302490234375, 0.32763671875, 0.03594970703125, -0.010162353515625, -0.31201171875, 0.05908203125, 0.27783203125, 0.3369140625, -0.71875, 0.61669921875, 0.09912109375, -0.1927490234375, 0.040771484375, -0.41796875, -0.0042724609375, -0.0294189453125, 0.373046875, -0.2998046875, 0.0086669921875, -0.1827392578125, -0.192138671875, 0.450927734375, -0.000640869140625, -0.3310546875, -0.6767578125, -0.34228515625, 0.268310546875, 0.0684814453125, -0.21630859375, 0.300537109375, -0.056640625, 0.294921875, -0.454345703125, -0.29541015625, 0.2108154296875, -0.3935546875, -0.420654296875, -0.6376953125, -0.6376953125, 0.01702880859375, 0.7509765625, 0.6533203125, -0.32568359375, 0.04656982421875, 0.6904296875, -0.5908203125, 0.2236328125, -0.69287109375, -0.65380859375, -0.5341796875, 0.163818359375, 0.0880126953125, -0.34375, -0.29443359375, 0.162109375, -0.1044921875, 0.14013671875, 0.45654296875, 0.0635986328125, 0.1165771484375, 0.030029296875, -0.5595703125, 0.08154296875, -0.0150146484375, -0.0633544921875, 0.79541015625, 0.083251953125, -0.33154296875, 0.46484375, -0.6416015625, 0.451171875, -0.2744140625, 0.5283203125, -0.1546630859375, 0.085693359375, 0.14306640625, -0.951171875, -0.145263671875, 0.0267333984375, 0.179931640625, -0.27490234375, -0.445556640625, -0.0325927734375, 1.0126953125, 0.37158203125, -0.61669921875, -0.218505859375, 0.67724609375, -0.14501953125, -0.18359375, -0.06011962890625, -0.46826171875, 0.23046875, -0.1553955078125, -0.4677734375, 0.453857421875, 0.4736328125, -0.2880859375, -0.54052734375, 0.078857421875, 0.0947265625, -0.12152099609375, 0.17529296875, -0.2958984375, -0.107177734375, 0.210693359375, 0.274169921875, 0.3505859375, 0.358642578125, 0.2197265625, -0.095703125, 0.16650390625, -0.0230712890625, 0.34912109375, 0.32470703125, 0.45361328125, -0.300537109375, -0.0869140625, 0.1783447265625, -0.29052734375, -0.02734375, -0.25634765625, 0.31591796875, 0.454833984375, -0.2435302734375, 0.48193359375, -2.140625, 0.50146484375, 0.424560546875, 0.11865234375, -0.4287109375, -0.192138671875, 0.107421875, -0.148193359375, 0.0552978515625, 0.25048828125, 0.447021484375, -0.169921875, 0.74658203125, -0.364990234375, 0.0162353515625, 0.66015625, 0.2327880859375, 0.2744140625, -0.21142578125, -0.3818359375, 0.2022705078125, 0.45166015625, 0.1773681640625, -0.072265625, -0.3056640625, 0.309814453125, 0.1103515625, 0.9599609375, 0.2734375, 0.351318359375, -0.278564453125, 0.373779296875, 0.213623046875, -0.00787353515625, 0.7060546875, -0.6552734375, -0.2998046875, 0.44384765625, -0.6123046875, -0.0438232421875, 0.7861328125, -0.440185546875, -0.232666015625, -0.0849609375, -0.43701171875, -0.16650390625, 0.2015380859375, -0.9775390625, -0.0616455078125, 0.06463623046875, -0.0289306640625, 0.6806640625, -0.30224609375, 0.39892578125, -0.298583984375, -0.2108154296875, -0.2437744140625, -0.07177734375, 0.047607421875, -0.10205078125, 0.15234375, 0.0052490234375, 0.113525390625, 0.29052734375, -0.098388671875, -0.1201171875, -0.390380859375, 0.014312744140625, 0.4169921875, 1.21875, -0.153564453125, -0.263427734375, -0.56982421875, 0.5380859375, 0.29443359375, 0.0560302734375, 0.284912109375, -0.310302734375, -1.357421875, -0.05279541015625, 0.05517578125, -0.59033203125, 0.43994140625, 0.51318359375, 0.019775390625, 0.20556640625, -0.0938720703125, 0.08984375, 0.13330078125, 0.05364990234375, -0.54931640625, 0.35986328125, -0.1331787109375, -0.480712890625, 0.2386474609375, -0.1129150390625, -0.3427734375, -0.1749267578125, -0.18212890625, -0.1204833984375, -0.45458984375, 0.049072265625, 0.0157470703125, -0.4052734375, -0.53076171875, -0.180419921875, -0.50439453125, -0.16845703125, 0.28515625, -0.036865234375, -0.2666015625, -0.1875, -0.192626953125, -0.17041015625, 0.140380859375, 0.50244140625, 0.0623779296875, 0.1884765625, -0.68212890625, -0.3203125, -0.483642578125, -0.24853515625, 0.028564453125, 0.191162109375, 0.179443359375, 0.0439453125, 0.40087890625, -0.0284576416015625, 0.137939453125, -0.03564453125, -0.2120361328125, 0.393798828125, 0.14892578125, 0.59033203125, -0.148193359375, 0.037109375, -0.787109375, -0.06280517578125, 0.037841796875, 0.187255859375, -0.25048828125, -0.07415771484375, -0.54150390625, 0.28125, -1.435546875, 0.1610107421875, -0.11480712890625, 1.248046875, -0.1263427734375, 0.41748046875, 0.3427734375, 0.397705078125, -0.2005615234375, 0.002197265625, 0.28564453125, -0.0404052734375, 0.40673828125, 0.55126953125, -0.59765625, -0.2236328125, -0.22021484375, -0.1427001953125, -0.025146484375, 0.175048828125, 0.43798828125, 0.08837890625, -0.634765625, 0.2427978515625, -0.2265625, -0.047760009765625, -0.7265625, 0.8212890625, -0.040283203125, -0.0208740234375, 0.0296630859375, -0.7353515625, 0.453125, 0.262451171875, 0.498046875, -0.0965576171875, -0.701171875, 0.54541015625, 0.40185546875, -0.0904541015625, 1.10546875, -0.0205078125, -0.1004638671875, -0.734375, -0.146728515625, -0.216796875, 0.0030517578125, -0.010986328125, 0.15380859375, 0.45068359375, -0.17333984375, 0.10107421875, 0.0411376953125, -0.53759765625, -0.275634765625, 0.21533203125, -0.46875, -0.27392578125, 0.10089111328125, -0.366943359375, -0.0548095703125, -0.0986328125, -0.021240234375, 0.83837890625, 0.08465576171875, -0.2548828125, -0.140625, 0.1871337890625, 0.05999755859375, -0.00592041015625, 0.236572265625, -0.1201171875, 0.267822265625, 0.537109375, -0.4033203125, 0.202880859375, -0.349853515625, -0.83154296875, 0.1409912109375, 0.272216796875, 0.1572265625, 0.03125, 0.07861328125, 0.08740234375, 0.110595703125, -0.188232421875, 0.409912109375, 0.6943359375]}, {"frame_number": 24, "vector": [-0.20361328125, 0.26708984375, -0.048828125, -0.17041015625, -0.6162109375, 0.322509765625, 0.0723876953125, -0.445556640625, -0.094482421875, -0.059051513671875, -0.082763671875, 0.05908203125, 0.333984375, 0.23193359375, 0.0079345703125, 0.28271484375, 0.132568359375, -0.16943359375, -0.047576904296875, -0.4921875, -0.576171875, -6.171875, -0.271484375, -0.71337890625, 0.10205078125, -0.1168212890625, -0.1572265625, -0.8671875, 0.473876953125, -0.203857421875, -0.9580078125, -0.5986328125, -0.712890625, 0.05780029296875, -0.1185302734375, 0.05938720703125, 0.3759765625, -0.64208984375, -0.07049560546875, -0.43994140625, -0.06781005859375, 0.202880859375, 0.16015625, 0.021240234375, 0.947265625, -0.357666015625, -0.38134765625, -0.1744384765625, -0.5908203125, -0.16015625, 0.0133514404296875, -0.272216796875, 0.08612060546875, 0.49267578125, 0.20751953125, 0.04046630859375, 0.404296875, -0.27294921875, 0.26611328125, 0.06036376953125, -0.04296875, 0.09912109375, 0.265380859375, 0.16943359375, -0.318603515625, -0.3642578125, -0.4912109375, 0.03271484375, 0.264404296875, 0.1585693359375, 0.8984375, -0.0126953125, -0.021240234375, -0.020263671875, -0.40869140625, 0.81494140625, -0.337646484375, -0.44775390625, 0.220947265625, -0.71923828125, 1.1884765625, 0.322265625, 0.58935546875, 0.47119140625, -0.27392578125, 0.057373046875, 0.1900634765625, -0.008758544921875, 0.1923828125, -0.6767578125, -0.1151123046875, -0.0750732421875, 0.0263671875, -0.135986328125, 0.328857421875, 0.380859375, 0.0675048828125, 0.46728515625, 0.0604248046875, -0.6416015625, 0.7744140625, 0.30419921875, -0.40625, -0.64306640625, 0.04400634765625, -0.068115234375, 0.08367919921875, -0.1065673828125, -0.1690673828125, -0.0732421875, -0.09686279296875, 0.128662109375, 0.1016845703125, -0.492431640625, 0.103759765625, -0.541015625, -0.155517578125, -0.0899658203125, 0.326416015625, 0.53759765625, 0.0814208984375, -0.242431640625, -0.027587890625, -0.6708984375, -0.0535888671875, 0.35595703125, -0.1630859375, 0.376953125, 0.1488037109375, 0.2978515625, -0.006591796875, -0.4609375, -0.02490234375, -0.0181884765625, 0.529296875, -0.0413818359375, 0.116943359375, 0.79345703125, 0.5361328125, 0.049560546875, 0.053466796875, 0.323974609375, 0.1544189453125, -0.148681640625, -0.2333984375, 0.06787109375, -0.06475830078125, -0.412841796875, 0.2191162109375, 1.0546875, -0.1448974609375, -1.560546875, 0.1151123046875, -0.4638671875, 0.392578125, -0.321533203125, -0.198974609375, 1.005859375, 0.3291015625, -0.46728515625, -0.29638671875, 0.22900390625, 0.419677734375, 0.412353515625, -0.3095703125, -0.247314453125, 0.0015869140625, 0.64990234375, 0.03076171875, 0.0411376953125, 0.1480712890625, 0.4677734375, -0.1917724609375, 0.63818359375, -0.2705078125, 1.0556640625, -1.0791015625, 0.127197265625, -0.0140380859375, -0.077880859375, 0.0283203125, 0.09765625, -0.1856689453125, 0.03448486328125, 0.0042724609375, 0.02008056640625, -0.01513671875, 0.1785888671875, -0.0789794921875, 0.0994873046875, 0.2369384765625, 0.1759033203125, 0.476806640625, -0.39697265625, 0.55908203125, -0.204833984375, 0.45166015625, -0.197021484375, -0.26220703125, 1.228515625, 0.5107421875, -0.03369140625, -0.1693115234375, 0.08251953125, -0.0465087890625, 0.0755615234375, 0.58984375, 0.00396728515625, -0.568359375, -0.1802978515625, 0.1788330078125, -0.023681640625, 0.23193359375, -0.84130859375, -0.544921875, -0.53271484375, 0.48583984375, 1.00390625, -0.0272216796875, -0.45263671875, -0.459228515625, -0.52490234375, -0.1107177734375, -0.587890625, -0.14599609375, 0.226806640625, -0.146728515625, -0.2412109375, 0.71337890625, 0.0277099609375, -0.03607177734375, 0.261474609375, 0.07763671875, 0.35400390625, -0.41845703125, -0.5283203125, 0.7333984375, -0.07275390625, -0.9755859375, -0.08251953125, 0.69970703125, 0.260986328125, -0.0628662109375, 0.190185546875, -0.111572265625, 0.962890625, -0.1978759765625, -0.002197265625, -0.269287109375, -0.111328125, 0.0140380859375, -0.05291748046875, 0.2998046875, 0.007080078125, -0.71337890625, -0.05322265625, -0.136962890625, -0.04443359375, 0.4267578125, 0.447265625, -0.0220947265625, -0.260009765625, -0.042236328125, -0.013427734375, -0.263671875, -0.0643310546875, -0.7939453125, -0.296875, 0.55859375, -0.406982421875, -0.009002685546875, 0.45751953125, 0.365234375, 0.256591796875, -0.2109375, -0.724609375, -0.1759033203125, 0.1649169921875, -0.347412109375, -0.3828125, -0.2432861328125, 0.6953125, -0.381591796875, 0.152587890625, 0.42822265625, 0.280029296875, -0.140625, 0.5849609375, -0.5654296875, 1.15625, -0.026947021484375, 0.08258056640625, 0.0528564453125, -0.44140625, 0.0504150390625, 0.0968017578125, 0.124755859375, 0.35302734375, 0.6201171875, -0.298095703125, 0.0165252685546875, 0.73583984375, -0.0594482421875, -0.035400390625, -0.326416015625, 0.37841796875, 0.363037109375, -0.65869140625, 0.11199951171875, -0.08203125, 0.365478515625, -0.362060546875, -0.00634765625, -0.282958984375, -0.3603515625, 0.31494140625, 0.01995849609375, -0.0013427734375, -0.263671875, 0.14111328125, -0.94677734375, 0.110595703125, 0.06451416015625, -0.44091796875, -0.89111328125, 0.54443359375, -0.39453125, 0.285888671875, -0.587890625, -0.07525634765625, -0.5419921875, -0.07177734375, -0.1409912109375, -0.5107421875, 0.1251220703125, 0.282958984375, 0.152587890625, -0.587890625, 0.03265380859375, -0.006591796875, -0.7568359375, 0.119140625, -0.327880859375, -0.6435546875, -0.587890625, 0.130126953125, -0.44921875, 0.39794921875, 0.158935546875, 0.012939453125, -0.300048828125, -0.2919921875, 0.311279296875, 0.185546875, 0.386962890625, 0.33544921875, 0.2440185546875, -0.1888427734375, -0.31787109375, -0.196533203125, -0.19384765625, -0.10052490234375, 0.5146484375, 0.26708984375, -0.5791015625, -0.462158203125, -0.009765625, 0.2406005859375, 0.059173583984375, -0.14404296875, -0.006591796875, 0.2239990234375, -0.064697265625, 0.32177734375, 0.31201171875, -0.08154296875, 0.3701171875, -0.200439453125, 0.11328125, -0.556640625, -0.03271484375, -0.352294921875, 0.13134765625, 0.70458984375, 0.164794921875, -0.0584716796875, -0.04254150390625, -0.144775390625, -0.197998046875, 0.4892578125, -0.658203125, -0.104736328125, 0.32080078125, 0.255859375, 1.001953125, 0.625, -0.077880859375, -0.7041015625, -0.094970703125, -0.3662109375, 0.25244140625, 0.241455078125, -0.0966796875, -0.39990234375, -0.154052734375, 0.026611328125, -0.3740234375, 0.03106689453125, 0.36328125, -0.105224609375, 0.206298828125, -0.90283203125, -0.7041015625, -0.24853515625, -0.3408203125, 0.505859375, 0.759765625, -0.259765625, 1.166015625, -0.28076171875, 0.179931640625, -0.06793212890625, -0.214111328125, -0.083984375, 0.0101318359375, -0.51171875, -1.361328125, -0.0513916015625, 0.1400146484375, 0.244140625, -0.010009765625, -6.1875, -0.37744140625, -0.044677734375, -0.81884765625, -0.237548828125, -0.1202392578125, -1.1953125, -0.73291015625, -0.264404296875, 0.0133056640625, 0.6640625, 0.65234375, -0.20703125, -0.0379638671875, 0.140625, 0.21044921875, -0.0147705078125, -0.103271484375, 0.19189453125, -0.025146484375, 0.199462890625, -0.09033203125, 0.1806640625, 0.416748046875, -0.279052734375, -0.1793212890625, 0.30712890625, -0.5712890625, -0.615234375, -0.087646484375, 0.054443359375, -0.35009765625, -0.294921875, -0.4521484375, -0.0592041015625, 0.60107421875, 0.767578125, -0.26806640625, -0.41357421875, 0.2392578125, -0.97119140625, 0.316162109375, -0.0068359375, -0.0733642578125, 0.215087890625, 0.343505859375, -0.126708984375, -0.7294921875, -0.74853515625, 0.12548828125, 0.1363525390625, -1.0634765625, 0.022705078125, 0.325927734375, -0.001708984375, 0.23583984375, -0.078857421875, 0.106689453125, 0.1748046875, 0.24462890625, 0.5390625, 0.32666015625, 0.447509765625, -0.43994140625, 0.017822265625, 0.062744140625, -0.01611328125, 0.13525390625, 0.01800537109375, -0.05224609375, -0.031829833984375, 0.67138671875, -0.23095703125, -0.31396484375, -0.061767578125, -0.04931640625, 0.0684814453125, -0.12939453125, 0.120849609375, 0.2200927734375, 0.05908203125, 0.102294921875, 0.310546875, -0.221923828125, -0.232177734375, -0.101318359375, -0.39306640625, -0.61328125, -0.10546875, 0.67724609375, 0.2080078125, -0.2266845703125, 0.0125732421875, 0.248779296875, 1.2568359375, 0.17138671875, 0.19873046875, 0.09820556640625, 0.15576171875, -0.378173828125, -0.466064453125, 0.047607421875, -0.05029296875, 0.38037109375, 0.21484375, 0.015625, 0.0235595703125, 0.41064453125, -0.0003662109375, 0.6572265625, 0.10150146484375, 0.372802734375, 0.529296875, -0.1265869140625, 0.474365234375, -1.5224609375, -0.33984375, -0.30615234375, 0.1483154296875, -0.02587890625, -0.273193359375, 0.7294921875, 0.3125, -0.0538330078125, 0.027587890625, 0.287841796875, -0.0821533203125, -0.277587890625, -0.99658203125, -0.475830078125, -0.289794921875, -0.405029296875, -0.87841796875, 0.365478515625, 0.3095703125, 0.443359375, -0.34619140625, -0.1495361328125, 0.0848388671875, -0.0301513671875, 0.35107421875, -0.216796875, 0.1312255859375, 0.197265625, 0.444091796875, -0.388671875, 0.192138671875, -0.3251953125, -0.46923828125, -0.3837890625, 0.1689453125, -0.218994140625, -0.1163330078125, -0.01068115234375, -0.118408203125, 0.03466796875, 0.4794921875, -0.260498046875, 0.22998046875, 0.3984375, 0.0158843994140625, -0.1845703125, 0.03570556640625, 0.0843505859375, 0.086181640625, -0.130615234375, -0.2353515625, -1.5, -0.43408203125, 0.2138671875, -0.0411376953125, -0.33544921875, -0.4169921875, -0.1912841796875, -0.06982421875, -0.11126708984375, -0.281494140625, -0.1859130859375, -0.865234375, 0.189453125, 0.34228515625, -0.1900634765625, -0.18701171875, -0.1099853515625, -0.5595703125, 0.275390625, 0.274658203125, 0.00152587890625, -0.20703125, 0.083251953125, -0.100830078125, 0.08056640625, 0.0804443359375, -0.287353515625, -0.0034332275390625, 0.026123046875, 0.382568359375, 0.215087890625, -0.087158203125, 0.19384765625, 0.003997802734375, -0.418212890625, 0.6591796875, 0.1170654296875, -0.193115234375, -0.269775390625, 1.3291015625, 0.12353515625, -0.296630859375, 0.09423828125, -0.44384765625, 0.354736328125, -0.20947265625, -0.194091796875, 0.343994140625, 0.160888671875, 0.03179931640625, 0.262939453125, 0.07000732421875, 0.307373046875, -0.303955078125, 0.343505859375, 0.2412109375, -0.685546875, 0.18408203125, 0.258544921875, -0.0989990234375, -0.24951171875, -0.17431640625, -0.33935546875, -0.83544921875, 0.00689697265625, -0.050537109375, -0.07745361328125, -0.4033203125, 0.048828125, -0.046142578125, 0.45361328125, -0.4443359375, -1.33984375, -0.0596923828125, -0.112060546875, 0.18994140625, 0.2210693359375, -0.09130859375, 0.392578125, 0.0975341796875, -0.05029296875, 0.299072265625, 0.0223388671875, 0.23486328125, 0.44970703125, 0.025634765625, 0.107421875, -0.1337890625, 0.29736328125, 0.199951171875, -0.4150390625, -0.603515625, -0.1502685546875, -0.357421875, 0.03662109375, 0.2353515625, -0.299560546875, 0.85693359375, -0.097900390625, 0.1268310546875, -0.84228515625, 0.001953125, -0.01470947265625, 0.1234130859375, -0.5966796875, 0.298095703125, -0.5634765625, 0.3828125, -0.0540771484375, 0.230224609375, -0.0997314453125, -0.50048828125, -0.69140625, 0.1256103515625, -0.473388671875, 0.48388671875, -0.220947265625, 0.9599609375, -0.294921875, 0.0927734375, -0.172607421875, -0.060546875, -0.41357421875, 0.041259765625, 1.419921875, 0.080322265625, 0.299072265625, -0.016845703125, 0.399169921875, -1.16015625, -0.217529296875, 0.47509765625, 0.02288818359375, -0.25, 0.5322265625, 0.193359375, 0.090576171875, 0.1875, 0.099853515625, 0.204345703125, -0.03369140625, 0.50341796875, -0.5146484375, 0.057373046875, -0.58642578125, 0.0546875, 0.0975341796875, -0.5009765625, -0.064208984375, -0.039306640625, -0.0784912109375, 0.0623779296875, -0.83447265625, -0.23046875, 0.271484375, -0.041259765625, 0.2705078125, -0.01763916015625, -0.114501953125, -0.79541015625, 0.06201171875, 0.265869140625, 0.0264892578125, 0.0819091796875, -0.2481689453125, -0.55517578125, 0.5478515625, -0.429931640625, 0.02978515625, -0.6279296875, 0.03997802734375, -0.489013671875, 0.10162353515625, -0.0048828125, 0.05419921875, 0.0328369140625, 0.2156982421875, -2.26171875, 0.296142578125, -0.1552734375, 0.2734375, 0.6376953125, -0.0164794921875, -0.375, -0.180419921875, 0.1640625, -0.12890625, 0.2333984375, 0.053955078125, 0.1663818359375, 0.583984375, 0.3984375, 0.339111328125, 0.017333984375, 0.24755859375, 0.6083984375, -0.29541015625, 0.32666015625, 0.0416259765625, -0.0228271484375, -0.33642578125, 0.07421875, 0.2333984375, 0.33837890625, -0.759765625, 0.6142578125, 0.0921630859375, -0.25244140625, 0.064453125, -0.451171875, 0.0277099609375, -0.05859375, 0.3984375, -0.307373046875, -0.002655029296875, -0.18310546875, -0.21337890625, 0.446044921875, 0.0009765625, -0.330078125, -0.6689453125, -0.357666015625, 0.2802734375, 0.097412109375, -0.218505859375, 0.357666015625, -0.021484375, 0.272705078125, -0.436279296875, -0.2734375, 0.2115478515625, -0.3974609375, -0.4775390625, -0.65869140625, -0.6123046875, 0.00787353515625, 0.74609375, 0.61767578125, -0.312255859375, 0.0213623046875, 0.703125, -0.57470703125, 0.239990234375, -0.666015625, -0.67431640625, -0.5546875, 0.175537109375, 0.0755615234375, -0.335693359375, -0.283203125, 0.18359375, -0.12890625, 0.125, 0.455078125, 0.04937744140625, 0.0992431640625, 0.016204833984375, -0.546875, 0.06396484375, -0.01995849609375, -0.065185546875, 0.7919921875, 0.05615234375, -0.296875, 0.50732421875, -0.67724609375, 0.450927734375, -0.294677734375, 0.59765625, -0.1427001953125, 0.0791015625, 0.1495361328125, -0.974609375, -0.143798828125, 0.0128173828125, 0.195556640625, -0.31982421875, -0.459716796875, -0.0013427734375, 1.0078125, 0.33203125, -0.6005859375, -0.2421875, 0.72021484375, -0.150390625, -0.205322265625, -0.0946044921875, -0.455078125, 0.2088623046875, -0.1668701171875, -0.4697265625, 0.447998046875, 0.46044921875, -0.287841796875, -0.5322265625, 0.0924072265625, 0.0765380859375, -0.1199951171875, 0.168701171875, -0.2457275390625, -0.10736083984375, 0.2047119140625, 0.274169921875, 0.3798828125, 0.36767578125, 0.188720703125, -0.096923828125, 0.170654296875, -0.0386962890625, 0.35205078125, 0.319580078125, 0.494140625, -0.28076171875, -0.07177734375, 0.1944580078125, -0.276123046875, -0.0068359375, -0.22509765625, 0.31396484375, 0.407958984375, -0.2861328125, 0.521484375, -2.142578125, 0.552734375, 0.44287109375, 0.097412109375, -0.438232421875, -0.168212890625, 0.107177734375, -0.150634765625, 0.058837890625, 0.26904296875, 0.402099609375, -0.11865234375, 0.7841796875, -0.350830078125, 0.00390625, 0.61279296875, 0.2275390625, 0.26806640625, -0.20947265625, -0.39892578125, 0.1866455078125, 0.46240234375, 0.1912841796875, -0.09716796875, -0.199951171875, 0.279052734375, 0.10443115234375, 0.95751953125, 0.289306640625, 0.326171875, -0.3173828125, 0.359619140625, 0.22509765625, -0.026519775390625, 0.6640625, -0.68994140625, -0.341796875, 0.44921875, -0.630859375, -0.0074462890625, 0.78125, -0.4501953125, -0.196044921875, -0.0810546875, -0.392578125, -0.197509765625, 0.2073974609375, -0.9775390625, -0.0775146484375, 0.0885009765625, -0.0120849609375, 0.6962890625, -0.2744140625, 0.38232421875, -0.298095703125, -0.187744140625, -0.2392578125, -0.064453125, 0.045654296875, -0.08740234375, 0.104736328125, 0.029296875, 0.0780029296875, 0.313232421875, -0.116943359375, -0.133544921875, -0.384765625, -0.018280029296875, 0.46533203125, 1.193359375, -0.1612548828125, -0.261474609375, -0.6357421875, 0.5087890625, 0.3505859375, 0.0604248046875, 0.310546875, -0.2469482421875, -1.357421875, -0.092041015625, 0.076416015625, -0.607421875, 0.4208984375, 0.5390625, 0.02630615234375, 0.19921875, -0.11785888671875, 0.0810546875, 0.159912109375, 0.038330078125, -0.5185546875, 0.35400390625, -0.15234375, -0.51416015625, 0.2415771484375, -0.135009765625, -0.35107421875, -0.15966796875, -0.170654296875, -0.1549072265625, -0.39111328125, 0.060546875, 0.020263671875, -0.384033203125, -0.54443359375, -0.186279296875, -0.5263671875, -0.116455078125, 0.25244140625, -0.07373046875, -0.246337890625, -0.21142578125, -0.197509765625, -0.182373046875, 0.1748046875, 0.493408203125, 0.09814453125, 0.1873779296875, -0.69580078125, -0.33740234375, -0.49755859375, -0.262939453125, 0.0306396484375, 0.166748046875, 0.195556640625, 0.035400390625, 0.421142578125, -0.066162109375, 0.1304931640625, -0.01885986328125, -0.1961669921875, 0.395751953125, 0.142333984375, 0.587890625, -0.1153564453125, 0.05712890625, -0.8203125, -0.07342529296875, 0.069091796875, 0.1778564453125, -0.2454833984375, -0.04461669921875, -0.54443359375, 0.30224609375, -1.4482421875, 0.1400146484375, -0.11468505859375, 1.2314453125, -0.11822509765625, 0.3720703125, 0.321044921875, 0.437255859375, -0.1845703125, -0.0225830078125, 0.29443359375, -0.0279541015625, 0.443359375, 0.5390625, -0.61669921875, -0.2236328125, -0.197021484375, -0.12139892578125, -0.04119873046875, 0.185546875, 0.438232421875, 0.0869140625, -0.630859375, 0.2354736328125, -0.2109375, -0.053619384765625, -0.71875, 0.83056640625, -0.023681640625, -0.054931640625, 0.018798828125, -0.74609375, 0.44873046875, 0.25439453125, 0.5068359375, -0.088134765625, -0.708984375, 0.59765625, 0.34619140625, -0.079345703125, 1.109375, 0.009521484375, -0.10888671875, -0.7216796875, -0.08453369140625, -0.247314453125, 0.002685546875, 0.01171875, 0.1390380859375, 0.43603515625, -0.180419921875, 0.1158447265625, -0.00152587890625, -0.49755859375, -0.281494140625, 0.243896484375, -0.48388671875, -0.2978515625, 0.09381103515625, -0.3720703125, -0.069580078125, -0.1287841796875, -0.04730224609375, 0.8427734375, 0.10284423828125, -0.2374267578125, -0.129638671875, 0.1884765625, 0.09954833984375, -0.0096435546875, 0.194091796875, -0.122314453125, 0.29638671875, 0.5576171875, -0.3876953125, 0.199951171875, -0.3623046875, -0.8046875, 0.107421875, 0.2802734375, 0.1298828125, 0.00250244140625, 0.036376953125, 0.10009765625, 0.148193359375, -0.169189453125, 0.427490234375, 0.63330078125]}, {"frame_number": 25, "vector": [-0.231201171875, 0.261474609375, -0.0623779296875, -0.14404296875, -0.60595703125, 0.361083984375, 0.01116943359375, -0.447509765625, -0.099609375, -0.08160400390625, -0.104736328125, 0.0625, 0.283203125, 0.214599609375, -0.0179443359375, 0.25732421875, 0.09527587890625, -0.16015625, -0.0772705078125, -0.4794921875, -0.52880859375, -6.2578125, -0.283447265625, -0.66259765625, 0.0953369140625, -0.0714111328125, -0.17041015625, -0.8837890625, 0.46044921875, -0.213134765625, -0.9736328125, -0.59130859375, -0.69482421875, 0.04205322265625, -0.09527587890625, 0.0662841796875, 0.384521484375, -0.69580078125, -0.06689453125, -0.475341796875, -0.02093505859375, 0.2359619140625, 0.15087890625, 0.0103759765625, 1.0087890625, -0.376220703125, -0.371826171875, -0.1639404296875, -0.55615234375, -0.14013671875, 0.041534423828125, -0.2900390625, 0.1085205078125, 0.473388671875, 0.21142578125, -0.0382080078125, 0.420166015625, -0.288330078125, 0.278076171875, 0.04296875, -0.061279296875, 0.15673828125, 0.26123046875, 0.17333984375, -0.288330078125, -0.29150390625, -0.481689453125, 0.01513671875, 0.2880859375, 0.1767578125, 0.9658203125, -0.023193359375, -0.026123046875, -0.0421142578125, -0.377685546875, 0.796875, -0.3310546875, -0.42724609375, 0.246826171875, -0.71728515625, 1.177734375, 0.286376953125, 0.6328125, 0.53125, -0.2431640625, 0.027099609375, 0.2291259765625, -0.01910400390625, 0.216796875, -0.6728515625, -0.0933837890625, -0.101318359375, 0.0316162109375, -0.119384765625, 0.33837890625, 0.353515625, 0.0791015625, 0.47265625, 0.06597900390625, -0.6669921875, 0.75390625, 0.3564453125, -0.39013671875, -0.6767578125, 0.0380859375, -0.0838623046875, 0.1220703125, -0.1201171875, -0.187255859375, -0.070068359375, -0.115234375, 0.1220703125, 0.047821044921875, -0.447021484375, 0.10009765625, -0.5517578125, -0.181396484375, -0.10986328125, 0.343994140625, 0.51904296875, 0.08154296875, -0.244140625, -0.0272216796875, -0.662109375, -0.04052734375, 0.3681640625, -0.1268310546875, 0.36962890625, 0.177978515625, 0.27197265625, -0.0125732421875, -0.47802734375, -0.0169677734375, 0.001495361328125, 0.54931640625, -0.0650634765625, 0.086669921875, 0.76904296875, 0.5390625, 0.047607421875, 0.0401611328125, 0.32177734375, 0.198486328125, -0.0987548828125, -0.271484375, 0.100341796875, -0.029296875, -0.418701171875, 0.21484375, 1.0859375, -0.140625, -1.5947265625, 0.084716796875, -0.44970703125, 0.39990234375, -0.317138671875, -0.18896484375, 1.021484375, 0.35595703125, -0.486328125, -0.2410888671875, 0.2197265625, 0.4306640625, 0.33935546875, -0.3125, -0.239013671875, -0.0179443359375, 0.623046875, 0.035888671875, 0.0224609375, 0.20361328125, 0.40478515625, -0.19775390625, 0.60498046875, -0.186767578125, 1.056640625, -1.14453125, 0.1287841796875, 0.0062255859375, -0.10009765625, 0.037353515625, 0.164306640625, -0.2021484375, -0.035888671875, 0.0587158203125, 0.0333251953125, -0.03271484375, 0.1610107421875, -0.0968017578125, 0.1634521484375, 0.185302734375, 0.164306640625, 0.470703125, -0.3828125, 0.50341796875, -0.2213134765625, 0.4736328125, -0.25390625, -0.24169921875, 1.146484375, 0.5126953125, -0.04736328125, -0.213134765625, 0.065673828125, -0.02252197265625, 0.121337890625, 0.64990234375, 0.01824951171875, -0.55908203125, -0.184326171875, 0.17919921875, -0.078369140625, 0.23828125, -0.81689453125, -0.50341796875, -0.515625, 0.5234375, 1.0, 0.0147705078125, -0.46533203125, -0.472900390625, -0.55615234375, -0.1278076171875, -0.56396484375, -0.152587890625, 0.2060546875, -0.1275634765625, -0.265625, 0.76513671875, 0.005859375, -0.0479736328125, 0.27001953125, 0.044189453125, 0.344970703125, -0.334228515625, -0.56494140625, 0.75732421875, -0.1290283203125, -0.916015625, -0.039306640625, 0.75244140625, 0.281494140625, -0.0765380859375, 0.239990234375, -0.14306640625, 0.9794921875, -0.213134765625, -0.0205078125, -0.3095703125, -0.132080078125, 0.001953125, -0.0272216796875, 0.277587890625, -0.047119140625, -0.72314453125, -0.060791015625, -0.097412109375, -0.074462890625, 0.418701171875, 0.4736328125, 0.01446533203125, -0.2254638671875, -0.0458984375, -0.0267333984375, -0.306396484375, -0.023193359375, -0.76953125, -0.285888671875, 0.51953125, -0.421875, 0.04864501953125, 0.4609375, 0.357666015625, 0.237060546875, -0.21044921875, -0.70263671875, -0.1376953125, 0.15478515625, -0.349365234375, -0.376708984375, -0.263916015625, 0.6787109375, -0.412109375, 0.21630859375, 0.421630859375, 0.3056640625, -0.1512451171875, 0.57568359375, -0.60009765625, 1.1796875, -0.05194091796875, 0.034820556640625, -0.0084228515625, -0.444580078125, 0.0572509765625, 0.096435546875, 0.125732421875, 0.41064453125, 0.59228515625, -0.26953125, 0.00984954833984375, 0.74072265625, -0.027740478515625, -0.046875, -0.339111328125, 0.4169921875, 0.38720703125, -0.67138671875, 0.117431640625, -0.091796875, 0.3359375, -0.38232421875, 0.030029296875, -0.275390625, -0.351318359375, 0.334716796875, 0.0482177734375, 0.0238037109375, -0.255126953125, 0.124267578125, -0.9326171875, 0.11474609375, 0.011474609375, -0.41650390625, -0.896484375, 0.5791015625, -0.413330078125, 0.27490234375, -0.59375, -0.0821533203125, -0.5458984375, -0.074462890625, -0.166015625, -0.440673828125, 0.136962890625, 0.280517578125, 0.142822265625, -0.587890625, 0.0433349609375, -0.0048828125, -0.76123046875, 0.1365966796875, -0.326416015625, -0.66796875, -0.5908203125, 0.175048828125, -0.46630859375, 0.38427734375, 0.150390625, -0.0147705078125, -0.328369140625, -0.3701171875, 0.30126953125, 0.138916015625, 0.3955078125, 0.31298828125, 0.2337646484375, -0.163330078125, -0.290283203125, -0.250732421875, -0.1771240234375, -0.0675048828125, 0.48583984375, 0.20751953125, -0.568359375, -0.463134765625, -0.041748046875, 0.26220703125, 0.057098388671875, -0.150390625, -0.0408935546875, 0.230712890625, -0.047607421875, 0.3662109375, 0.306396484375, -0.112060546875, 0.3662109375, -0.296875, 0.08349609375, -0.56103515625, -0.04052734375, -0.29248046875, 0.1124267578125, 0.74755859375, 0.150390625, -0.05908203125, -0.03900146484375, -0.150634765625, -0.2587890625, 0.50439453125, -0.634765625, -0.12353515625, 0.285888671875, 0.236083984375, 1.03125, 0.64453125, -0.078369140625, -0.6875, -0.0667724609375, -0.32177734375, 0.279052734375, 0.2374267578125, -0.099365234375, -0.41357421875, -0.16064453125, 0.039306640625, -0.39697265625, 0.03778076171875, 0.330810546875, -0.1611328125, 0.245361328125, -0.87548828125, -0.7333984375, -0.2607421875, -0.363037109375, 0.5224609375, 0.74267578125, -0.2666015625, 1.205078125, -0.291015625, 0.1748046875, -0.10357666015625, -0.265625, -0.070068359375, -0.0159912109375, -0.484375, -1.361328125, -0.0572509765625, 0.1575927734375, 0.239013671875, 0.00360107421875, -6.28125, -0.38037109375, -0.0107421875, -0.7666015625, -0.28466796875, -0.11083984375, -1.2138671875, -0.728515625, -0.275390625, -0.01385498046875, 0.63330078125, 0.64697265625, -0.218017578125, 0.005126953125, 0.25537109375, 0.276611328125, 0.0306396484375, -0.127197265625, 0.18408203125, -0.0419921875, 0.199951171875, -0.09075927734375, 0.1796875, 0.410888671875, -0.2296142578125, -0.1689453125, 0.29736328125, -0.626953125, -0.642578125, -0.11865234375, 0.097412109375, -0.335205078125, -0.315673828125, -0.482421875, -0.001953125, 0.5693359375, 0.78515625, -0.31396484375, -0.3818359375, 0.243896484375, -1.009765625, 0.2724609375, -0.0240478515625, -0.032562255859375, 0.2034912109375, 0.3955078125, -0.0921630859375, -0.7216796875, -0.81005859375, 0.09820556640625, 0.1134033203125, -1.03125, -0.039794921875, 0.332763671875, -0.0284423828125, 0.2147216796875, -0.06005859375, 0.0814208984375, 0.1898193359375, 0.243896484375, 0.51318359375, 0.359619140625, 0.445556640625, -0.42626953125, 0.0439453125, 0.07159423828125, -0.022216796875, 0.13525390625, 0.033203125, -0.0467529296875, -0.0721435546875, 0.634765625, -0.1737060546875, -0.2958984375, -0.069091796875, -0.044677734375, 0.1160888671875, -0.16796875, 0.143310546875, 0.23681640625, 0.043212890625, 0.102783203125, 0.337890625, -0.22412109375, -0.183349609375, -0.10400390625, -0.377197265625, -0.61328125, -0.119873046875, 0.6591796875, 0.151611328125, -0.228759765625, 0.01171875, 0.23388671875, 1.267578125, 0.189208984375, 0.2021484375, 0.0933837890625, 0.1405029296875, -0.38134765625, -0.51171875, 0.05517578125, -0.06640625, 0.38525390625, 0.238037109375, 0.03271484375, 0.01422119140625, 0.447265625, 0.034912109375, 0.677734375, 0.1064453125, 0.353515625, 0.47265625, -0.16845703125, 0.50830078125, -1.5439453125, -0.34423828125, -0.28564453125, 0.138427734375, -0.01904296875, -0.293212890625, 0.73828125, 0.294921875, -0.05303955078125, 0.03515625, 0.299560546875, -0.0494384765625, -0.28369140625, -1.005859375, -0.4921875, -0.28173828125, -0.3984375, -0.9072265625, 0.366943359375, 0.362060546875, 0.465087890625, -0.38916015625, -0.138671875, 0.0950927734375, -0.0338134765625, 0.360595703125, -0.248046875, 0.084716796875, 0.1849365234375, 0.428466796875, -0.41650390625, 0.216796875, -0.26025390625, -0.5048828125, -0.35400390625, 0.123046875, -0.16650390625, -0.145263671875, -0.043212890625, -0.08544921875, 0.02392578125, 0.463623046875, -0.2509765625, 0.2113037109375, 0.4345703125, 0.0186614990234375, -0.2005615234375, 0.045562744140625, 0.093994140625, 0.109130859375, -0.12841796875, -0.248291015625, -1.470703125, -0.44140625, 0.2352294921875, -0.080322265625, -0.32666015625, -0.39697265625, -0.23876953125, -0.064453125, -0.130615234375, -0.244873046875, -0.236328125, -0.88671875, 0.1378173828125, 0.33349609375, -0.2183837890625, -0.18798828125, -0.123779296875, -0.513671875, 0.27197265625, 0.25927734375, -0.018798828125, -0.187744140625, 0.0379638671875, -0.05615234375, 0.0330810546875, 0.080078125, -0.3076171875, 0.001434326171875, 0.034912109375, 0.341796875, 0.210693359375, -0.0897216796875, 0.186767578125, -0.02435302734375, -0.408203125, 0.654296875, 0.0889892578125, -0.168701171875, -0.3046875, 1.3583984375, 0.120849609375, -0.297607421875, 0.08984375, -0.46435546875, 0.39208984375, -0.19921875, -0.193359375, 0.343017578125, 0.193603515625, 0.019287109375, 0.251220703125, 0.0733642578125, 0.314697265625, -0.305908203125, 0.334228515625, 0.266357421875, -0.67724609375, 0.23046875, 0.2353515625, -0.09423828125, -0.2578125, -0.1695556640625, -0.340576171875, -0.83349609375, -0.0062255859375, -0.05206298828125, -0.07537841796875, -0.35546875, 0.076416015625, -0.047607421875, 0.439453125, -0.40869140625, -1.4072265625, -0.069580078125, -0.062255859375, 0.18115234375, 0.2274169921875, -0.09527587890625, 0.3720703125, 0.07421875, -0.0159912109375, 0.368408203125, 0.0189208984375, 0.25830078125, 0.42138671875, 0.03564453125, 0.153076171875, -0.2078857421875, 0.2958984375, 0.233642578125, -0.40185546875, -0.599609375, -0.177978515625, -0.36962890625, 0.06298828125, 0.227783203125, -0.282958984375, 0.78515625, -0.0994873046875, 0.09326171875, -0.86865234375, 0.028564453125, 0.01806640625, 0.1314697265625, -0.5810546875, 0.3154296875, -0.54052734375, 0.333984375, -0.05364990234375, 0.2373046875, -0.103515625, -0.47998046875, -0.65625, 0.078369140625, -0.49951171875, 0.43603515625, -0.263916015625, 0.931640625, -0.319580078125, 0.1031494140625, -0.132080078125, -0.039794921875, -0.3857421875, 0.101806640625, 1.4892578125, 0.09698486328125, 0.2783203125, -0.03076171875, 0.364501953125, -1.033203125, -0.2332763671875, 0.525390625, 0.08111572265625, -0.2152099609375, 0.568359375, 0.119384765625, 0.08447265625, 0.19287109375, 0.05267333984375, 0.23828125, -0.017822265625, 0.4873046875, -0.51123046875, 0.0771484375, -0.5947265625, 0.0498046875, 0.0994873046875, -0.49560546875, -0.0762939453125, -0.03125, -0.116455078125, 0.1007080078125, -0.8193359375, -0.2459716796875, 0.3076171875, -0.11376953125, 0.266357421875, -0.01593017578125, -0.0843505859375, -0.8037109375, 0.07086181640625, 0.31298828125, 0.0426025390625, 0.064697265625, -0.269775390625, -0.53955078125, 0.5380859375, -0.456787109375, 0.0753173828125, -0.66064453125, 0.05072021484375, -0.482177734375, 0.10113525390625, 0.05810546875, 0.08966064453125, 0.037445068359375, 0.20068359375, -2.29296875, 0.314208984375, -0.131103515625, 0.251708984375, 0.671875, -0.0152587890625, -0.390625, -0.18994140625, 0.16552734375, -0.1123046875, 0.230224609375, 0.067138671875, 0.1689453125, 0.537109375, 0.377685546875, 0.2998046875, 0.027099609375, 0.2103271484375, 0.61083984375, -0.287841796875, 0.31640625, 0.050048828125, -0.008087158203125, -0.36865234375, 0.127197265625, 0.232666015625, 0.35888671875, -0.74609375, 0.603515625, 0.1390380859375, -0.1422119140625, 0.07275390625, -0.447265625, 0.0350341796875, -0.048095703125, 0.3974609375, -0.3505859375, 0.01416015625, -0.1839599609375, -0.213134765625, 0.403076171875, 0.0079345703125, -0.343017578125, -0.67431640625, -0.372802734375, 0.270263671875, 0.0677490234375, -0.21484375, 0.35693359375, -0.0704345703125, 0.270263671875, -0.469482421875, -0.251953125, 0.161865234375, -0.329833984375, -0.501953125, -0.65380859375, -0.6328125, 0.020233154296875, 0.7568359375, 0.6708984375, -0.297119140625, 0.00732421875, 0.64501953125, -0.61376953125, 0.19921875, -0.7099609375, -0.67333984375, -0.52734375, 0.160888671875, 0.0738525390625, -0.3125, -0.237060546875, 0.1309814453125, -0.17333984375, 0.15185546875, 0.465087890625, 0.057281494140625, 0.1326904296875, 0.05718994140625, -0.53857421875, 0.08740234375, -0.06689453125, -0.06982421875, 0.7509765625, 0.069580078125, -0.317138671875, 0.501953125, -0.6591796875, 0.4345703125, -0.303955078125, 0.6005859375, -0.1455078125, 0.054443359375, 0.1654052734375, -1.0048828125, -0.133056640625, 0.0323486328125, 0.21044921875, -0.302001953125, -0.475830078125, 0.015625, 1.0537109375, 0.3154296875, -0.6015625, -0.267333984375, 0.7421875, -0.1783447265625, -0.247802734375, -0.06768798828125, -0.4296875, 0.2093505859375, -0.210205078125, -0.49560546875, 0.44482421875, 0.460693359375, -0.2880859375, -0.57763671875, 0.0675048828125, 0.1041259765625, -0.126708984375, 0.123046875, -0.279296875, -0.1156005859375, 0.21435546875, 0.266845703125, 0.34423828125, 0.362060546875, 0.14404296875, -0.088623046875, 0.17822265625, -0.03631591796875, 0.34033203125, 0.343505859375, 0.49609375, -0.249755859375, -0.0679931640625, 0.153076171875, -0.23876953125, -0.0693359375, -0.240966796875, 0.310546875, 0.436767578125, -0.28369140625, 0.50830078125, -2.140625, 0.54296875, 0.447021484375, 0.1513671875, -0.5, -0.1754150390625, 0.1116943359375, -0.1728515625, 0.09619140625, 0.2236328125, 0.436279296875, -0.140869140625, 0.791015625, -0.328125, 0.00390625, 0.65087890625, 0.2486572265625, 0.26171875, -0.240234375, -0.401123046875, 0.152587890625, 0.45849609375, 0.1651611328125, -0.093505859375, -0.2276611328125, 0.300048828125, 0.111572265625, 0.9775390625, 0.30517578125, 0.324462890625, -0.341796875, 0.378662109375, 0.18310546875, -0.0699462890625, 0.7421875, -0.7138671875, -0.3408203125, 0.48193359375, -0.61767578125, 0.0196533203125, 0.787109375, -0.462890625, -0.2109375, -0.097412109375, -0.363525390625, -0.17724609375, 0.1614990234375, -0.955078125, -0.0933837890625, 0.074951171875, -0.0133056640625, 0.67578125, -0.29150390625, 0.416015625, -0.29052734375, -0.2093505859375, -0.188720703125, -0.0732421875, 0.04248046875, -0.1580810546875, 0.1065673828125, 0.056121826171875, 0.0948486328125, 0.295654296875, -0.12353515625, -0.1318359375, -0.390869140625, -0.026580810546875, 0.5068359375, 1.30859375, -0.141845703125, -0.227783203125, -0.56396484375, 0.5087890625, 0.35205078125, 0.049560546875, 0.34716796875, -0.2392578125, -1.349609375, -0.1383056640625, 0.078369140625, -0.6103515625, 0.3984375, 0.5517578125, 0.0084228515625, 0.16748046875, -0.11090087890625, 0.0826416015625, 0.156005859375, 0.03826904296875, -0.53173828125, 0.324462890625, -0.12158203125, -0.479736328125, 0.242919921875, -0.1473388671875, -0.34375, -0.135986328125, -0.0947265625, -0.189453125, -0.39794921875, 0.081787109375, -0.0025634765625, -0.421630859375, -0.46728515625, -0.184814453125, -0.51025390625, -0.121337890625, 0.270263671875, -0.02099609375, -0.238037109375, -0.220458984375, -0.217529296875, -0.1978759765625, 0.1025390625, 0.48095703125, 0.06640625, 0.169677734375, -0.71533203125, -0.320068359375, -0.497802734375, -0.1939697265625, 0.02734375, 0.216552734375, 0.194091796875, 0.03363037109375, 0.3984375, -0.08868408203125, 0.0850830078125, -0.00396728515625, -0.2352294921875, 0.39453125, 0.162353515625, 0.60595703125, -0.1055908203125, 0.05078125, -0.8154296875, -0.08538818359375, 0.0252685546875, 0.1817626953125, -0.230712890625, -0.05426025390625, -0.51318359375, 0.3330078125, -1.423828125, 0.0904541015625, -0.0791015625, 1.2001953125, -0.12115478515625, 0.40087890625, 0.348876953125, 0.403076171875, -0.1802978515625, -0.07122802734375, 0.295654296875, -0.052001953125, 0.4541015625, 0.560546875, -0.60205078125, -0.224365234375, -0.1612548828125, -0.143798828125, -0.071044921875, 0.164794921875, 0.443359375, 0.04693603515625, -0.5859375, 0.2493896484375, -0.205322265625, -0.0098876953125, -0.6787109375, 0.84619140625, -0.002685546875, -0.0589599609375, -0.012451171875, -0.724609375, 0.467529296875, 0.26220703125, 0.5009765625, -0.1302490234375, -0.74169921875, 0.61669921875, 0.2685546875, -0.1258544921875, 1.0947265625, -0.0145263671875, -0.098876953125, -0.6943359375, -0.09576416015625, -0.284423828125, 0.0084228515625, 0.035888671875, 0.143310546875, 0.509765625, -0.142822265625, 0.140380859375, 0.0054931640625, -0.52587890625, -0.318359375, 0.278564453125, -0.48681640625, -0.295654296875, 0.08349609375, -0.372314453125, -0.041015625, -0.10040283203125, -0.019287109375, 0.8740234375, 0.10546875, -0.2890625, -0.10986328125, 0.173095703125, 0.08740234375, 0.01324462890625, 0.24951171875, -0.1134033203125, 0.281494140625, 0.5595703125, -0.388671875, 0.207275390625, -0.32470703125, -0.837890625, 0.14892578125, 0.292236328125, 0.091552734375, 0.0091552734375, 0.0452880859375, 0.092041015625, 0.123779296875, -0.180908203125, 0.430419921875, 0.623046875]}]] \ No newline at end of file diff --git a/min_repro.py b/min_repro.py new file mode 100644 index 0000000..a6b86fd --- /dev/null +++ b/min_repro.py @@ -0,0 +1,336 @@ +import io + +import tensorrt as trt +import torch +import torch.nn as nn +import torch.nn.functional as F + + +class AttentionUsingScaledDotProduct(nn.Module): + """ + An alternative implementation of the Attention layer using `F.scaled_dot_product_attention`, which is ~50% faster, + but doesn't compile correctly when using TensorRT v10. + """ + + def __init__( + self, + dim, + num_heads=8, + qkv_bias=False, + qk_scale=None, + attn_drop=0.0, + proj_drop=0.0, + attn_head_dim=None, + ): + super().__init__() + self.num_heads = num_heads + head_dim = dim // num_heads + if attn_head_dim is not None: + head_dim = attn_head_dim + all_head_dim = head_dim * self.num_heads + self.scale = qk_scale or head_dim**-0.5 + + self.qkv = nn.Linear(dim, all_head_dim * 3, bias=False) + if qkv_bias: + self.q_bias = nn.Parameter(torch.zeros(all_head_dim)) + self.v_bias = nn.Parameter(torch.zeros(all_head_dim)) + else: + self.q_bias = None + self.v_bias = None + + self.attn_drop = nn.Dropout(attn_drop) + self.proj = nn.Linear(all_head_dim, dim) + self.proj_drop = nn.Dropout(proj_drop) + + def forward(self, x): + B, N, C = x.shape + qkv_bias = None + if self.q_bias is not None: + qkv_bias = torch.cat( + ( + self.q_bias, + torch.zeros_like(self.v_bias, requires_grad=False), + self.v_bias, + ) + ) + qkv = F.linear(input=x, weight=self.qkv.weight, bias=qkv_bias) + qkv = qkv.reshape(B, N, 3, self.num_heads, -1).permute(2, 0, 3, 1, 4) + q, k, v = qkv.unbind(0) + + x = F.scaled_dot_product_attention( + q, + k, + v, + dropout_p=self.attn_drop.p if self.training else 0.0, + scale=self.scale, + ) + + x = x.transpose(1, 2).reshape(B, N, -1) + + x = self.proj(x) + x = self.proj_drop(x) + return x + + +class ExplicitAttention(nn.Module): + """ + The explicit, original version of the Attention layer from the VideoMAEv2 codebase. + """ + + def __init__( + self, + dim, + num_heads=8, + qkv_bias=False, + qk_scale=None, + attn_drop=0.0, + proj_drop=0.0, + attn_head_dim=None, + ): + super().__init__() + self.num_heads = num_heads + head_dim = dim // num_heads + if attn_head_dim is not None: + head_dim = attn_head_dim + all_head_dim = head_dim * self.num_heads + self.scale = qk_scale or head_dim**-0.5 + + self.qkv = nn.Linear(dim, all_head_dim * 3, bias=False) + if qkv_bias: + self.q_bias = nn.Parameter(torch.zeros(all_head_dim)) + self.v_bias = nn.Parameter(torch.zeros(all_head_dim)) + else: + self.q_bias = None + self.v_bias = None + + self.attn_drop = nn.Dropout(attn_drop) + self.proj = nn.Linear(all_head_dim, dim) + self.proj_drop = nn.Dropout(proj_drop) + + def forward(self, x): + B, N, C = x.shape + qkv_bias = None + if self.q_bias is not None: + qkv_bias = torch.cat( + ( + self.q_bias, + torch.zeros_like(self.v_bias, requires_grad=False), + self.v_bias, + ) + ) + qkv = F.linear(input=x, weight=self.qkv.weight, bias=qkv_bias) + qkv = qkv.reshape(B, N, 3, self.num_heads, -1).permute(2, 0, 3, 1, 4) + q, k, v = qkv.unbind(0) + + q = q * self.scale + attn = q @ k.transpose(-2, -1) + + attn = attn.softmax(dim=-1) + attn = self.attn_drop(attn) + + x = (attn @ v).transpose(1, 2).reshape(B, N, -1) + + x = self.proj(x) + x = self.proj_drop(x) + + return x + + +class AttentionUsingMHAForward(nn.Module): + def __init__( + self, + dim, + num_heads=8, + qkv_bias=False, + qk_scale=None, + attn_drop=0.0, + proj_drop=0.0, + attn_head_dim=None, + ): + super().__init__() + self.num_heads = num_heads + head_dim = dim // num_heads + if attn_head_dim is not None: + head_dim = attn_head_dim + all_head_dim = head_dim * self.num_heads + self.scale = qk_scale or head_dim**-0.5 + + self.qkv = nn.Linear(dim, all_head_dim * 3, bias=False) + if qkv_bias: + self.q_bias = nn.Parameter(torch.zeros(all_head_dim)) + self.v_bias = nn.Parameter(torch.zeros(all_head_dim)) + else: + self.q_bias = None + self.v_bias = None + + self.attn_drop = nn.Dropout(attn_drop) + self.proj = nn.Linear(all_head_dim, dim) + self.proj_drop = nn.Dropout(proj_drop) + + def forward(self, x): + B, N, C = x.shape + qkv_bias = None + if self.q_bias is not None: + qkv_bias = torch.cat( + ( + self.q_bias, + torch.zeros_like(self.v_bias, requires_grad=False), + self.v_bias, + ) + ) + qkv = F.linear(input=x, weight=self.qkv.weight, bias=qkv_bias) + qkv = qkv.reshape(B, N, 3, self.num_heads, -1).permute(2, 0, 3, 1, 4) + q, k, v = qkv.unbind(0) + + # MHA expects [sequence, batch, embed_dim]. + x_t = x.transpose(0, 1) # => [N, B, C] + + attn_out, _ = F.multi_head_attention_forward( + x_t, + x_t, + x_t, + embed_dim_to_check=C, + num_heads=self.num_heads, + # Since use_separate_proj_weight=False (default), then according to the docs: + # "in_proj_weight will be used, which is a combination of q_proj_weight, k_proj_weight, v_proj_weight." + in_proj_weight=self.qkv.weight, + in_proj_bias=qkv_bias, + bias_k=None, + bias_v=None, + add_zero_attn=False, + dropout_p=self.attn_drop.p, + out_proj_weight=self.proj.weight, + out_proj_bias=self.proj.bias, + training=self.training, + key_padding_mask=None, + need_weights=False, + attn_mask=None, + ) + + # Transpose back to [B, N, C]. + x = attn_out.transpose(0, 1) + + return x + + +def onnx_to_trt(onnx_bytes: bytes) -> bytes: + TRT_LOGGER = trt.Logger(trt.Logger.INFO) + builder = trt.Builder(TRT_LOGGER) + + network = builder.create_network() + parser = trt.OnnxParser(network, TRT_LOGGER) + + parser.parse(onnx_bytes) + + config = builder.create_builder_config() + config.builder_optimization_level = 0 + + engine = builder.build_serialized_network(network, config) + + return engine + + +def build_trt_module(model, x): + onnx_bytes = io.BytesIO() + + torch.onnx.export( + model, + (x,), + onnx_bytes, + export_params=True, + opset_version=17, + do_constant_folding=True, + input_names=["x"], + output_names=["y"], + ) + + trt_engine = onnx_to_trt(onnx_bytes.getvalue()) + return trt_engine + + + +#@torch.inference_mode() +#def main(): +with torch.no_grad(): + torch.manual_seed(0) + + EMB_DIM = 384 + x = torch.rand((6, 1568, EMB_DIM)) + + explicit_attention = ExplicitAttention(EMB_DIM) + sdpa = AttentionUsingScaledDotProduct(EMB_DIM) + mha_fwd = AttentionUsingMHAForward(EMB_DIM) + + # Use the same params for all. + sdpa.load_state_dict(explicit_attention.state_dict()) + mha_fwd.load_state_dict(explicit_attention.state_dict()) + + sdpa_torch_y = sdpa(x) + explicit_attention_torch_y = explicit_attention(x) + mha_fwd_torch_y = mha_fwd(x) + + print( + "Torch: [explicit<->sdpa] Is allclose?", + sdpa_torch_y.allclose(explicit_attention_torch_y, atol=0.0001), + ) + print( + "Torch: [explicit<->mha_fwd] Is allclose?", + mha_fwd_torch_y.allclose(explicit_attention_torch_y, atol=0.0001), + ) + print( + "Torch: [explicit<->sdpa] Total difference:", + (sdpa_torch_y - explicit_attention_torch_y).abs().sum(), + ) + print( + "Torch: [explicit<->mha_fwd] Total difference:", + (mha_fwd_torch_y - explicit_attention_torch_y).abs().sum(), + ) + assert sdpa_torch_y.allclose(explicit_attention_torch_y, atol=0.0001), "Precheck" + assert mha_fwd_torch_y.allclose(explicit_attention_torch_y, atol=0.0001), "Precheck" +# %% + + explicit_attention_trt = build_trt_module(explicit_attention, x) + with open('explicit_attention_trt.trt','wb') as ea: + ea.write(explicit_attention_trt) + + sdpa_trt_model = build_trt_module(sdpa, x) + with open('sdpa_trt.trt','wb') as ea: + ea.write(sdpa_trt_model) + + mha_fwd_trt_model = build_trt_module(mha_fwd, x) + with open('mha_trt.trt','wb') as ea: + ea.write(mha_fwd_trt_model) +# %% + + + +# %% + explicit_attention_y = explicit_attention_trt(x.cuda()) + sdpa_y = sdpa_trt_model(x.cuda()) + mha_fwd_y = mha_fwd_trt_model(x.cuda()) + + print( + "TRT: [explicit<->sdpa] Is allclose?", + sdpa_y.allclose(explicit_attention_y, atol=0.0001), + ) + print( + "TRT: [explicit<->sdpa] Total difference:", + (sdpa_y - explicit_attention_y).abs().sum(), + ) + + print( + "TRT: [explicit<->mha_fwd] Is allclose?", + mha_fwd_y.allclose(explicit_attention_y, atol=0.0001), + ) + print( + "TRT: [explicit<->mha_fwd] Total difference:", + (mha_fwd_y - explicit_attention_y).abs().sum(), + ) + + print("TRT: Explicit Attention:", explicit_attention_y[0, 0, :32]) + print("TRT: Scaled Dot Product Attention:", sdpa_y[0, 0, :32]) + print("TRT: MHA Forward:", mha_fwd_y[0, 0, :32]) + + +if __name__ == "__main__": + main() diff --git a/ml_run.py b/ml_run.py new file mode 100644 index 0000000..62e76dc --- /dev/null +++ b/ml_run.py @@ -0,0 +1,334 @@ +import sys + +sys.path.insert(0, "/home/thebears/source/models/yolov7") +import time +from datetime import datetime +import cv2 +import numpy as np +from pymediainfo import MediaInfo +import inspect +import open_clip +import sys +import torch +import yaml +from models.experimental import attempt_load +from utils.general import check_img_size, non_max_suppression +from torchvision import transforms + +device = torch.device("cuda") + +pretrained_name = "webli" +#model_name = "ViT-SO400M-16-SigLIP2-512" +#model_name = 'ViT-SO400M-14-SigLIP-384' + +clip_model, _, clip_preprocess_og = open_clip.create_model_and_transforms( + model_name, pretrained=pretrained_name +) +tokenizer = open_clip.get_tokenizer('hf-hub:timm/'+model_name) +labels_list = ["A bird with a brown head and black body", "A bird with a black head and black body"] +text = tokenizer(labels_list, context_length=clip_model.context_length) + +import torch.nn.functional as F +with torch.no_grad(): + text_features = clip_model.encode_text(text).detach().cpu() + text_features = F.normalize(text_features, dim=-1).detach().cpu() +# %% + +clip_model = clip_model.half().to(device) +clip_dtype = next(clip_model.parameters()).dtype +clip_img_size = clip_preprocess_og.transforms[0].size +_ = clip_model.encode_image( + torch.rand(1, 3, *clip_img_size, dtype=clip_dtype, device=device) +) +clip_preprocess = transforms.Compose([clip_preprocess_og.transforms[x] for x in [0, 3]]) + +det_root_path = "/home/thebears/source/model_weights" +det_model_weights_root = os.path.join(det_root_path, "yolov7") +det_model_weights_path = os.path.join(det_model_weights_root, "best.pt") +det_data_yaml_path = os.path.join(det_model_weights_root, "inaturalist.yaml") +det_model = attempt_load(det_model_weights_path, map_location=device) +det_model = det_model.half().to(device) + +det_dtype = next(det_model.parameters()).dtype +det_imgsz = 1280 +det_stride = int(det_model.stride.max()) +det_imgsz = check_img_size(det_imgsz, s=det_stride) +_ = det_model(torch.zeros(1, 3, det_imgsz, det_imgsz, dtype=det_dtype).to(device)) + +with open(det_data_yaml_path, "r") as ff: + det_model_info = yaml.safe_load(ff) + det_labels = det_model_info["names"] + + + + +array_score = clip_array +frame_numbers = [x[0] for x in array_score] +frame_values = [x[1] for x in array_score] +frame_as_tensor = ( + torch.from_numpy(np.stack(frame_values)[:, :, :, 0:3]) + .to(torch.float16) + .to(device) + .permute([0, 3, 1, 2]) +) +def score_frames_det(array_score): + frame_numbers = [x[0] for x in array_score] + frame_values = [x[1] for x in array_score] + frame_as_tensor = ( + torch.from_numpy(np.stack(frame_values)[:, :, :, 0:3]) + .to(torch.float16) + .to(device) + .permute([0, 3, 1, 2]) + ) + + with torch.no_grad(): + frame_for_model = det_vid_preprocess(frame_as_tensor).div(255)[:,[2,1,0],:,:] + det_preds = det_model(frame_for_model)[0] + det_pred_post_nms = non_max_suppression(det_preds,0.25, 0.5) + det_cpu_pred = [x.detach().cpu().numpy() for x in det_pred_post_nms] + +# frame_for_clip = clip_preprocess(frame_as_tensor[:,[0,1,2],:,:]) +# clip_pred = clip_model.encode_image(frame_for_clip).detach().cpu().numpy() + + return {"det": det_cpu_pred, "fr#": frame_numbers} + +def score_frames_clip(array_score): + frame_numbers = [x[0] for x in array_score] + frame_values = [x[1] for x in array_score] + frame_as_tensor = ( + torch.from_numpy(np.stack(frame_values)[:, :, :, 0:3]) + .to(torch.float16) + .to(device) + .permute([0, 3, 1, 2]) + ) + + with torch.no_grad(): +# frame_for_model = det_vid_preprocess(frame_as_tensor).div(255)[:,[2,1,0],:,:] +# det_preds = det_model(frame_for_model)[0] +# det_pred_post_nms = non_max_suppression(det_preds,0.25, 0.5) +# det_cpu_pred = [x.detach().cpu().numpy() for x in det_pred_post_nms] + + frame_for_clip = clip_preprocess(frame_as_tensor[:,[0,1,2],:,:]) + clip_pred = clip_model.encode_image(frame_for_clip).detach().cpu().numpy() + + + return {"clip": clip_pred, "fr#": frame_numbers} + + + + + +with torch.no_grad(): + frame_for_model = det_vid_preprocess(frame_as_tensor).div(255)[:,[2,1,0],:,:] + det_preds = det_model(frame_for_model)[0] + det_pred_post_nms = non_max_suppression(det_preds,0.25, 0.5) + det_cpu_pred = [x.detach().cpu().numpy() for x in det_pred_post_nms] + frame_for_clip = frame_as_tensor.div(255) + + frame_for_clip = clip_preprocess(frame_for_clip[:,(2,1,0),:,:]) + clip_pred = clip_model.encode_image(frame_for_clip).detach().cpu().numpy() + +score_result = {"det": det_cpu_pred, "clip": clip_pred, "fr#": frame_numbers} + + +clip_orin = F.normalize(torch.from_numpy(score_result['clip'])) +clip_tree = F.normalize(torch.from_numpy(saved_emb)) +print(np.dot(clip_tree, clip_orin.T)) + + +mvo = mean_vec_out[0] +ooo = frame_for_clip[0].cpu().numpy() + +plt.close('all') +fig = plt.figure() + +ax1 = fig.add_subplot(3,2,1) +ax1.imshow(mvo[0]) +ax2 = fig.add_subplot(3,2,2) +ax2.imshow(ooo[0]) +ax3 = fig.add_subplot(3,2,3) +ax3.imshow(mvo[1]) +ax4 = fig.add_subplot(3,2,4) +ax4.imshow(ooo[1]) +ax5 = fig.add_subplot(3,2,5) +ax5.imshow(mvo[2]) +ax6 = fig.add_subplot(3,2,6) +ax6.imshow(ooo[2]) +fig.show() + +# %% + + +raw_vec_out +mean_vec_out + +# %% +file_to_score = "/home/thebears/source/ml_code/short.mp4" +vec_file = '/home/thebears/source/ml_code/short.npz' +out = np.load(vec_file) + +mean_vec_path = '/home/thebears/source/ml_code/as_np_mean.npy' +mean_vec_out = np.load(mean_vec_path) + +raw_vec_path = '/home/thebears/source/ml_code/as_np_raw.npy' +raw_vec_out = np.load(raw_vec_path) + +saved_fr = list(out['frame_numbers']) +saved_emb = out['embeds'] +import numpy as np + + + +def get_video_info(file_path): + file_info = MediaInfo.parse(file_path) + video_info = None + frame_count = 0 + if len(file_info.video_tracks) > 0: video_info = file_info.video_tracks[0] + + video_info.frame_count = int(video_info.frame_count) + return video_info + + +video_info = get_video_info(file_to_score) +vid_decoder = "h264parse" +if video_info.format.lower() == "HEVC".lower(): + vid_decoder = "h265parse" + + +gst_cmd = "filesrc location={file_to_score} ! qtdemux name=demux demux.video_0 ! queue ! {vid_decoder} ! nvv4l2decoder ! nvvidconv ! videoscale method=1 add-borders=false ! video/x-raw,width=1280,height=1280 ! appsink sync=false".format( + file_to_score=file_to_score, vid_decoder=vid_decoder +) + +# gst_cmd = "filesrc location={file_to_score} ! qtdemux name=demux demux.video_0 ! queue ! {vid_decoder} ! nvv4l2decoder ! nvvidconv ! videoscale method=1 add-borders=false ! appsink sync=false".format(file_to_score=file_to_score, vid_decoder=vid_decoder) + +cap_handle = cv2.VideoCapture(gst_cmd, cv2.CAP_GSTREAMER) + +target_max = det_imgsz +vid_h = video_info.height +vid_w = video_info.width + +if vid_h > vid_w: + target_h = target_max + target_w = target_max * vid_w / vid_h +elif vid_h == vid_w: + target_h = target_max + target_w = target_max +elif vid_h < vid_w: + target_h = target_max * vid_h / vid_w + target_w = target_max + +target_h = int(target_h) +target_w = int(target_w) + +pad_amt = [None, None, None, None] +if target_w % det_stride != 0: + off = det_stride - target_w % det_stride + new_w = target_w + off + pad_diff = new_w - target_w + pad_left = round(pad_diff / 2) + pad_right = pad_diff - pad_left + pad_amt[0] = pad_left + pad_amt[2] = pad_right +else: + pad_amt[0] = 0 + pad_amt[2] = 0 + +if target_h % det_stride != 0: + off = det_stride - target_h % det_stride + new_h = target_h + off + pad_diff = new_h - target_h + pad_up = round(pad_diff / 2) + pad_down = pad_diff - pad_up + pad_amt[1] = pad_up + pad_amt[3] = pad_down +else: + pad_amt[1] = 0 + pad_amt[3] = 0 + + +det_vid_preprocess = transforms.Compose( + [transforms.Resize((target_h, target_w)), transforms.Pad(pad_amt, fill=127)] +) + +batch_size = 6 +clip_interval = 10 +array_score = list() +final_output = dict() +final_output["start_score_time"] = time.time() +final_output["num_frames"] = video_info.frame_count +st = time.time() +frame_numbers = list() +det_results = list() +clip_results = list() + +clip_array = list() + +for i in range(video_info.frame_count): + success, frame_matrix = cap_handle.read() + + clip_array.append((i, frame_matrix)) + + if not success: + break + + + array_score.append((i, frame_matrix)) + + if len(array_score) >= batch_size: + score_result = score_frames(array_score) + + + det_results.extend(score_result["det"]) + clip_results.extend(score_result["clip"]) + frame_numbers.extend(score_result["fr#"]) + array_score = list() + + if not (i % clip_interval): + print('do_clip') + +if len(array_score) > 0: + score_result = score_frames(array_score) + det_results.extend(score_result["det"]) + clip_results.extend(score_result["clip"]) + frame_numbers.extend(score_result["fr#"]) + +cap_handle.release() +et = time.time() + +final_output["end_score_time"] = time.time() +final_output["video"] = { + "w": vid_w, + "h": vid_h, + "path": file_to_score, + "target_w": target_w, + "target_h": target_h, + "pad_amt": pad_amt, +} + +try: + final_output['scoring_fps'] = final_output['num_frames']/ (final_output['end_score_time'] - final_output['start_score_time']) +except Exception as e: + pass + +final_output['scores'] = list() + +for frame_number, frame in zip(frame_numbers, det_results): + cframe_dict = dict() + cframe_dict['frame'] = frame_number + cframe_dict['score_number'] = frame_number + cframe_dict['detections'] = list() + + for det in frame: + data = dict() + data['coords'] = [float(x) for x in list(det[0:4])] + data['score'] = float(det[4]) + data['idx'] = int(det[5]) + + try: + data['name'] = det_labels[data['idx']] + except: + data['name'] = 'Code failed' + + cframe_dict['detections'].append(data) + + final_output['scores'].append(cframe_dict) diff --git a/orin.png b/orin.png new file mode 100644 index 0000000..a5334b8 Binary files /dev/null and b/orin.png differ diff --git a/report_dynamo_export.sarif b/report_dynamo_export.sarif new file mode 100644 index 0000000..35af130 --- /dev/null +++ b/report_dynamo_export.sarif @@ -0,0 +1,95638 @@ +{ + "runs":[ + { + "tool":{ + "driver":{ + "name":"torch.onnx.dynamo_export", + "contents":[ + "localizedData", + "nonLocalizedData" + ], + "language":"en-US", + "rules":[ + { + "id":"FXE0016", + "fullDescription":{ + "text":"This rule involves finding the list of OnnxFunction for the PyTorch operator overload in the ONNX registry. If the operator overload is not supported but its default overload is, a warning will be issued. If both the operator overload and its default overload are not supported, an error will be issued.", + "markdown":"The operator overload name serves the purpose of verifying whether a PyTorch operator is registered in the ONNX registry.\nIf it's not found, the dispatcher takes a fallback approach and tries to locate the default overload of the PyTorch\noperator in the registry. If even the default overload is absent, it signifies that the operator is officially unsupported.\n\nThere are three types of level that can be triggered in this rule:\n\n1. NOTE: The op overload is supported.\n2. WARNING: The op overload is not supported, but it's default overload is supported.\n3. ERROR: The op overload is not supported, and it's default overload is also not supported.\n\nHere are some suggestions based on the WARNING situation:\n\n1. If there are NO errors or mismatches in the results, it is safe to disregard this warning.\n2. If there are errors or mismatches in the results, it is recommended to:\n (a) Enable op_level_debugging to determine if the OnnxFunction might be incorrect.\n (b) Report the unsupported overload to the PyTorch-ONNX team.\n (c) Create/register a custom symbolic function to replace the default one.\n\nHere are some suggestions based on the ERROR situation:\n\n1. Report the unsupported operator to the PyTorch-ONNX team.\n2. Create/register a custom symbolic function to replace the default one.\n" + }, + "name":"find-operator-overloads-in-onnx-registry", + "shortDescription":{ + "text":"Find the list of OnnxFunction of the PyTorch operator in onnx registry." + } + }, + { + "id":"FXE0007", + "fullDescription":{ + "text":"Transforms graph from FX IR to ONNX IR.", + "markdown":"This diagnostic tracks the transformation process from an FX Graph (in FX IR) to an ONNX Graph (in ONNX IR).\n\n## Key Representations:\n\n- **FX Graph**: The graph in FX IR produced by dynamo or symbolic tracing.\n- **ONNX Graph**: The graph in ONNX IR and [operators](https://onnx.ai/onnx/operators/).\n\n## Additional Notes:\n\n- Prior to this transformation step, the FX graph undergoes preprocessing through multiple FX passes.\n To gain insight into these transformations, refer to diagnostic `FXE0010`.\n- To enable a detailed view of the graph transformation in progress within this diagnostic, switch to the DEBUG mode.\n\n - Set DiagnosticOptions.verbosity_level to logging.DEBUG.\n - Activate the environment variable TORCH_LOGS='onnx_diagnostics'.\n\n- For specific information related to node-level FX to ONNX transformations, explore the diagnostic `FXE0008`.\n" + }, + "name":"fx-graph-to-onnx", + "shortDescription":{ + "text":"Transforms graph from FX IR to ONNX IR." + } + }, + { + "id":"FXE0015", + "fullDescription":{ + "text":"Determine if type promotion is required for the FX node. Insert cast nodes if needed.", + "markdown":"This diagnostic monitors the node-level type promotion insertion process. In PyTorch, there is an automatic process called implicit type promotion,\nwhere the input types of an operator are promoted to a common type. The determination of the common type is based on the type promotion rule specific to each operator.\nTo learn more about PyTorch's type promotion rules, refer to the [elementwise_dtypes doc](https://github.com/pytorch/pytorch/blob/f044613f78df713fb57f70c608483c9f10ad332e/torch/_prims_common/__init__.py#L1252-L1335)\nand [torch._refs ops](https://github.com/pytorch/pytorch/blob/a475ea4542dfe961c9d097e33ab5041f61c8c17f/torch/_refs/__init__.py#L484).\n\nHowever, implicit type promotion is not supported in ONNX. Therefore, to replicate the PyTorch behavior, we need to explicitly insert cast nodes.\nThis diagnostic tracks the process of node-level type promotion insertion.\n\nThe type promotion rules used by this process can be found in `torch/onnx/_internal/fx/passes/type_promotion.py.`\nTo update or add new type promotion rules, please refer to the [Note: Update type promotion rule] section.\n" + }, + "name":"fx-node-insert-type-promotion", + "shortDescription":{ + "text":"Determine if type promotion is required for the FX node. Insert cast nodes if needed." + } + }, + { + "id":"FXE0008", + "fullDescription":{ + "text":"Transforms an FX node to an ONNX node.", + "markdown":"This diagnostic tracks the transformation process from an FX Node to ONNX [Operators](https://onnx.ai/onnx/operators/).\n\nThe process of converting FX Node to ONNX Node involves dealing with six distinct node types:\n 1. `placeholder`: Represents a module input, maps to an ONNX graph input.\n 2. `call_module`: Symbolizes a call to a submodule, maps to an ONNX\n 3. `call_method`: Symbolizes a method call. Not yet implemented.\n 4. `call_function`: Symbolizes a function call. [Core ATen](https://pytorch.org/docs/stable/ir.html#core-aten-ir) is expected\n as the function call target. The mapping from ATen to ONNX is implemented by [ONNXScript torchlib](https://github.com/microsoft/onnxscript/tree/main/onnxscript/function_libs/torch_lib/ops).\n This [guide](https://pytorch.org/docs/stable/onnx.html#onnx-script-functions) shows how to write and register a custom symbolic function for call_function FX node.\n 5. `get_attr`: Indicates an attribute access within the current module. Maps to an ONNX graph initializer.\n 6. `output`: Represents the module's output. Maps to an ONNX graph output.\n\nFor a granular understanding of how each node type is transformed, refer to the implementation details in `FxOnnxInterpreter`.\n" + }, + "name":"fx-node-to-onnx", + "shortDescription":{ + "text":"Transforms an FX node to an ONNX node." + } + }, + { + "id":"FXE0014", + "fullDescription":{ + "text":"Find the OnnxFunction that matches the input dtypes by comparing them with their opschemas. A warning will be issued if the matched OnnxFunction is not an exact match.", + "markdown":"When an ATen/Custom operator is registered and needs to be dispatched to an OnnxFunction, the input/attribute\ndtypes of the ATen/Custom operator are compared with the input/attribute dtypes of the OnnxFunction opschemas\nto find a match. However, if a perfect/exact match is not found, the dispatcher will attempt to find\nthe nearest match with the highest number of input/attribute dtypes matching the OnnxFunction opschemas, while\nissuing a warning.\n\nThere are two types of level that can be triggered in this rule:\n\n1. NOTE: A perfect match is found, and no warning is issued.\n2. WARNING: The matched OnnxFunction is not a perfect/exact match.\n\nHere are some suggestions based on the WARNING situation:\n\n1. If there are NO errors or mismatches in the results, it is safe to disregard this warning,\n as the definition of OnnxFunction schema is usually more stringent.\n2. If there are errors or mismatches in the results, it is recommended to:\n (a) Enable op_level_debugging to determine if the OnnxFunction might be incorrect.\n (b) Report the issue to the PyTorch-ONNX team.\n (c) Create/register a custom symbolic function to replace the default one.\n" + }, + "name":"find-opschema-matched-symbolic-function", + "shortDescription":{ + "text":"Find the OnnxFunction that matches the input/attribute dtypes by comparing them with their opschemas." + } + }, + { + "id":"FXE0010", + "fullDescription":{ + "text":"FX graph transformation during ONNX export before converting from FX IR to ONNX IR.", + "markdown":"This diagnostic tracks the FX passes executed during the ONNX export process prior\nto converting from FX IR (Intermediate Representation) to ONNX IR.\n\nUnder the scope of ONNX export, an FX pass refers to a specific transformation applied to the FX GraphModule.\nThe primary aim of these passes is to streamline the graph into a format that aligns more with the ONNX IR.\nMoreover, these passes work to substitute unsupported FX IR features with those recognized and endorsed by\nONNX IR. Common transformations include, but aren't limited to, decomposition, functionalization and\ntype promotion.\n\nFor those who are interested in a comprehensive log detailing the modifications made during these passes,\nthere are a couple of options:\n\n- Set DiagnosticOptions.verbosity_level to logging.DEBUG.\n- Activate the environment variable TORCH_LOGS='onnx_diagnostics'.\n\nHowever, it's noteworthy that by default, such detailed logging is turned off. The primary reason being\nits considerable impact on performance.\n\nFor an in-depth understanding of each specific pass, please refer to the directory: torch/onnx/_internal/fx/passes.\n" + }, + "name":"fx-pass", + "shortDescription":{ + "text":"FX graph transformation during ONNX export before converting from FX IR to ONNX IR." + } + } + ], + "version":"2.5.0a0+872d972e41.nv24.08" + } + }, + "language":"en-US", + "newlineSequences":[ + "\r\n", + "\n" + ], + "results":[ + { + "message":{ + "markdown":"Skipped p_trunk_pos_embed: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_pos_embed)[placeholder]:Tensor(f32[1, 1024, 1024])\n## Return values\nTensor(f32[1, 1024, 1024])", + "text":"Skipped p_trunk_pos_embed: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_latent: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_latent)[placeholder]:Tensor(f32[1, 1, 1024])\n## Return values\nTensor(f32[1, 1, 1024])", + "text":"Skipped p_trunk_attn_pool_latent: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_patch_embed_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_patch_embed_proj_weight)[placeholder]:Tensor(f32[1024, 3, 16, 16])\n## Return values\nTensor(f32[1024, 3, 16, 16])", + "text":"Skipped p_trunk_patch_embed_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_patch_embed_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_patch_embed_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_trunk_patch_embed_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___0___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___0___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___0___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___0___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___0___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___0___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___0___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___0___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___0___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___0___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___0___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___0___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___0___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___0___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___0___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___0___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___1___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___1___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___1___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___1___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___1___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___1___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___1___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___1___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___1___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___1___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___1___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___1___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___1___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___1___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___1___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___1___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___2___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___2___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___2___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___2___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___2___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___2___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___2___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___2___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___2___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___2___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___2___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___2___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___2___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___2___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___2___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___2___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___3___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___3___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___3___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___3___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___3___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___3___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___3___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___3___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___3___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___3___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___3___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___3___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___3___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___3___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___3___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___3___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___4___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___4___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___4___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___4___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___4___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___4___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___4___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___4___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___4___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___4___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___4___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___4___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___4___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___4___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___4___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___4___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___5___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___5___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___5___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___5___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___5___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___5___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___5___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___5___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___5___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___5___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___5___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___5___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___5___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___5___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___5___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___5___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___6___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___6___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___6___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___6___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___6___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___6___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___6___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___6___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___6___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___6___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___6___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___6___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___6___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___6___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___6___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___6___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___7___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___7___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___7___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___7___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___7___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___7___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___7___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___7___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___7___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___7___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___7___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___7___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___7___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___7___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___7___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___7___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___8___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___8___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___8___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___8___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___8___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___8___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___8___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___8___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___8___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___8___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___8___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___8___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___8___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___8___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___8___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___8___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___9___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___9___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___9___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___9___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___9___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___9___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___9___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___9___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___9___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___9___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___9___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___9___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___9___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___9___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___9___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___9___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___10___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___10___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___10___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___10___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___10___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___10___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___10___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___10___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___10___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___10___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___10___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___10___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___10___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___10___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___10___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___10___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___11___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___11___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___11___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___11___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___11___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___11___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___11___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___11___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___11___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___11___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___11___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___11___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___11___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___11___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___11___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___11___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___12___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___12___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___12___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___12___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___12___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___12___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___12___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___12___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___12___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___12___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___12___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___12___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___12___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___12___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___12___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___12___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___13___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___13___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___13___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___13___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___13___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___13___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___13___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___13___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___13___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___13___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___13___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___13___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___13___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___13___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___13___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___13___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___14___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___14___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___14___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___14___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___14___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___14___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___14___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___14___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___14___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___14___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___14___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___14___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___14___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___14___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___14___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___14___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___15___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___15___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___15___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___15___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___15___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___15___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___15___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___15___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___15___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___15___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___15___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___15___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___15___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___15___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___15___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___15___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___16___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___16___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___16___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___16___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___16___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___16___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___16___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___16___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___16___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___16___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___16___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___16___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___16___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___16___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___16___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___16___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___17___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___17___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___17___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___17___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___17___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___17___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___17___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___17___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___17___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___17___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___17___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___17___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___17___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___17___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___17___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___17___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___18___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___18___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___18___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___18___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___18___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___18___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___18___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___18___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___18___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___18___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___18___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___18___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___18___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___18___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___18___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___18___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___19___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___19___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___19___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___19___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___19___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___19___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___19___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___19___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___19___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___19___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___19___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___19___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___19___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___19___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___19___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___19___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___20___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___20___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___20___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___20___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___20___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___20___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___20___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___20___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___20___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___20___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___20___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___20___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___20___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___20___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___20___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___20___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___21___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___21___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___21___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___21___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___21___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___21___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___21___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___21___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___21___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___21___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___21___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___21___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___21___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___21___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___21___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___21___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___22___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___22___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___22___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___22___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___22___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___22___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___22___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___22___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___22___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___22___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___22___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___22___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___22___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___22___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___22___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___22___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___23___norm1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm1_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___23___norm1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___23___norm1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm1_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___23___norm1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___23___attn_qkv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n## Return values\nTensor(f32[3072, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___23___attn_qkv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___23___attn_qkv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n## Return values\nTensor(f32[3072])", + "text":"Skipped p_getattr_l__self___trunk_blocks___23___attn_qkv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___23___attn_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___23___attn_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___23___attn_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___23___attn_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___23___norm2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm2_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___23___norm2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___23___norm2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___23___norm2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_norm_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_norm_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_trunk_norm_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_norm_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_norm_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_trunk_norm_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_q_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_q_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_trunk_attn_pool_q_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_q_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_q_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_trunk_attn_pool_q_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_kv_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_kv_weight)[placeholder]:Tensor(f32[2048, 1024])\n## Return values\nTensor(f32[2048, 1024])", + "text":"Skipped p_trunk_attn_pool_kv_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_kv_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_kv_bias)[placeholder]:Tensor(f32[2048])\n## Return values\nTensor(f32[2048])", + "text":"Skipped p_trunk_attn_pool_kv_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_proj_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped p_trunk_attn_pool_proj_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_proj_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_proj_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_trunk_attn_pool_proj_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_norm_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_norm_weight)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_trunk_attn_pool_norm_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_norm_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_norm_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_trunk_attn_pool_norm_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_mlp_fc1_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped p_trunk_attn_pool_mlp_fc1_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_mlp_fc1_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n## Return values\nTensor(f32[4096])", + "text":"Skipped p_trunk_attn_pool_mlp_fc1_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_mlp_fc2_weight: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped p_trunk_attn_pool_mlp_fc2_weight: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped p_trunk_attn_pool_mlp_fc2_bias: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n## Return values\nTensor(f32[1024])", + "text":"Skipped p_trunk_attn_pool_mlp_fc2_bias: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped x: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(x)[placeholder]:Tensor(f32[4, 3, 512, 512])\n## Return values\nTensor(f32[4, 3, 512, 512])", + "text":"Skipped x: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.convolution.default)[call_function]:Tensor(f32[4, 1024, 32, 32]): Cannot find type promotion rule for op: aten.convolution.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.convolution.default)[call_function]:Tensor(f32[4, 1024, 32, 32])\n## Return values\nTensor(f32[4, 1024, 32, 32])", + "text":"Skipped for fx.Node(aten.convolution.default)[call_function]:Tensor(f32[4, 1024, 32, 32]): Cannot find type promotion rule for op: aten.convolution.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument transpose is not promoted. Already torch.float32.\nArgument p_trunk_pos_embed is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_1. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument clone is not promoted. Already torch.float32.\nArgument clone_1 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_1. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_8 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_2. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_1 is not promoted. Already torch.float32.\nArgument clone_3 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_2. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_3. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_2 is not promoted. Already torch.float32.\nArgument clone_4 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_3. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_1. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_18 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_1. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_4. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_3 is not promoted. Already torch.float32.\nArgument clone_6 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_4. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_5. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_4 is not promoted. Already torch.float32.\nArgument clone_7 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_5. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_2. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_28 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_2. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_6. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_5 is not promoted. Already torch.float32.\nArgument clone_9 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_6. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_7. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_6 is not promoted. Already torch.float32.\nArgument clone_10 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_7. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_3. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_38 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_3. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_8. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_7 is not promoted. Already torch.float32.\nArgument clone_12 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_8. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_9. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_8 is not promoted. Already torch.float32.\nArgument clone_13 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_9. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_4. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_48 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_4. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_10. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_9 is not promoted. Already torch.float32.\nArgument clone_15 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_10. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_11. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_10 is not promoted. Already torch.float32.\nArgument clone_16 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_11. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_5. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_58 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_5. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_12. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_11 is not promoted. Already torch.float32.\nArgument clone_18 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_12. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_13. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_12 is not promoted. Already torch.float32.\nArgument clone_19 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_13. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_6. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_68 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_6. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_14. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_13 is not promoted. Already torch.float32.\nArgument clone_21 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_14. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_15. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_14 is not promoted. Already torch.float32.\nArgument clone_22 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_15. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_7. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_78 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_7. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_16. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_15 is not promoted. Already torch.float32.\nArgument clone_24 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_16. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_17. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_16 is not promoted. Already torch.float32.\nArgument clone_25 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_17. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_8. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_88 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_8. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_18. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_17 is not promoted. Already torch.float32.\nArgument clone_27 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_18. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_19. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_18 is not promoted. Already torch.float32.\nArgument clone_28 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_19. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_9. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_98 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_9. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_20. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_19 is not promoted. Already torch.float32.\nArgument clone_30 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_20. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_21. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_20 is not promoted. Already torch.float32.\nArgument clone_31 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_21. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_10. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_108 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_10. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_22. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_21 is not promoted. Already torch.float32.\nArgument clone_33 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_22. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_23. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_22 is not promoted. Already torch.float32.\nArgument clone_34 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_23. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_11. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_118 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_11. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_24. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_23 is not promoted. Already torch.float32.\nArgument clone_36 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_24. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_25. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_24 is not promoted. Already torch.float32.\nArgument clone_37 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_25. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_12. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_128 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_12. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_26. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_25 is not promoted. Already torch.float32.\nArgument clone_39 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_26. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_27. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_26 is not promoted. Already torch.float32.\nArgument clone_40 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_27. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_13. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_138 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_13. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_28. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_27 is not promoted. Already torch.float32.\nArgument clone_42 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_28. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_29. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_28 is not promoted. Already torch.float32.\nArgument clone_43 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_29. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_14. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_148 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_14. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_30. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_29 is not promoted. Already torch.float32.\nArgument clone_45 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_30. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_31. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_30 is not promoted. Already torch.float32.\nArgument clone_46 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_31. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_15. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_158 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_15. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_32. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_31 is not promoted. Already torch.float32.\nArgument clone_48 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_32. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_33. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_32 is not promoted. Already torch.float32.\nArgument clone_49 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_33. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_16. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_168 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_16. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_34. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_33 is not promoted. Already torch.float32.\nArgument clone_51 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_34. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_35. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_34 is not promoted. Already torch.float32.\nArgument clone_52 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_35. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_17. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_178 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_17. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_36. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_35 is not promoted. Already torch.float32.\nArgument clone_54 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_36. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_37. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_36 is not promoted. Already torch.float32.\nArgument clone_55 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_37. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_18. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_188 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_18. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_38. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_37 is not promoted. Already torch.float32.\nArgument clone_57 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_38. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_39. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_38 is not promoted. Already torch.float32.\nArgument clone_58 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_39. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_19. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_198 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_19. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_40. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_39 is not promoted. Already torch.float32.\nArgument clone_60 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_40. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_41. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_40 is not promoted. Already torch.float32.\nArgument clone_61 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_41. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_20. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_208 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_20. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_42. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_41 is not promoted. Already torch.float32.\nArgument clone_63 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_42. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_43. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_42 is not promoted. Already torch.float32.\nArgument clone_64 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_43. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_21. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_218 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_21. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_44. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_43 is not promoted. Already torch.float32.\nArgument clone_66 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_44. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_45. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_44 is not promoted. Already torch.float32.\nArgument clone_67 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_45. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_22. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_228 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_22. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_46. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_45 is not promoted. Already torch.float32.\nArgument clone_69 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_46. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n## Return values\nTensor(f32[1024, 3072])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n## Return values\nTensor(f32[4096, 3072])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n## Return values\nTensor(f32[4, 1024, 3072])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n## Return values\nTensor(f32[4, 1024, 3, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n## Return values\nTensor(f32[3, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n## Return values\nTensor(f32[4, 1024, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_47. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_46 is not promoted. Already torch.float32.\nArgument clone_70 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_47. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_23. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_238 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Type promotion not needed for gelu_23. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096])\n## Return values\nTensor(f32[4, 1024, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096])\n## Return values\nTensor(f32[4096, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_48. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument add_47 is not promoted. Already torch.float32.\nArgument clone_72 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Type promotion not needed for add_48. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n## Return values\nTensor(f32[4, 1024, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1024, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.expand.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.expand.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.expand.default)[call_function]:Tensor(f32[4, 1, 1024])\n## Return values\nTensor(f32[4, 1, 1024])", + "text":"Skipped for fx.Node(aten.expand.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.expand.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024])\n## Return values\nTensor(f32[4, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.mm.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.mm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.mm.default)[call_function]:Tensor(f32[4, 1024])\n## Return values\nTensor(f32[4, 1024])", + "text":"Skipped for fx.Node(aten.mm.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.mm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 1024])\n## Return values\nTensor(f32[4, 1, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_49. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_242 is not promoted. Already torch.float32.\nArgument p_trunk_attn_pool_q_bias is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1, 1024])", + "text":"Type promotion not needed for add_49. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 16, 64])\n## Return values\nTensor(f32[4, 1, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 16, 1, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 16, 1, 64])\n## Return values\nTensor(f32[4, 16, 1, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 16, 1, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 2048]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 2048])\n## Return values\nTensor(f32[1024, 2048])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 2048]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 2048]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 2048])\n## Return values\nTensor(f32[4096, 2048])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 2048]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 2048]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 2048])\n## Return values\nTensor(f32[4, 1024, 2048])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 2048]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 2, 16, 64]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 2, 16, 64])\n## Return values\nTensor(f32[4, 1024, 2, 16, 64])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 2, 16, 64]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[2, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[2, 4, 16, 1024, 64])\n## Return values\nTensor(f32[2, 4, 16, 1024, 64])", + "text":"Skipped for fx.Node(aten.permute.default)[call_function]:Tensor(f32[2, 4, 16, 1024, 64]): Cannot find type promotion rule for op: aten.permute.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=2](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=2](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n## Return values\nList[length=2](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)", + "text":"Skipped for fx.Node(aten.unbind.int)[call_function]:List[length=2](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n): Cannot find type promotion rule for op: aten.unbind.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n## Return values\nTensor(f32[4, 16, 1024, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n## Return values\nTuple[length=4](\nTensor(f32[4, 16, 1, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)", + "text":"Skipped for fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n): Cannot find type promotion rule for op: aten._scaled_dot_product_efficient_attention.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1, 64]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1, 64])\n## Return values\nTensor(f32[4, 16, 1, 64])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 16, 1, 64]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1, 16, 64])\n## Return values\nTensor(f32[4, 1, 16, 64])", + "text":"Skipped for fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1, 16, 64]): Cannot find type promotion rule for op: aten.transpose.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 1024])\n## Return values\nTensor(f32[4, 1, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024])\n## Return values\nTensor(f32[4, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n## Return values\nTensor(f32[1024, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4, 1024])\n## Return values\nTensor(f32[4, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 1024])\n## Return values\nTensor(f32[4, 1, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1, 1024])\n## Return values\nTensor(f32[4, 1, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1, 1024]),\nTensor(f32[4, 1, 1]),\nTensor(f32[4, 1, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1, 1024]),\nTensor(f32[4, 1, 1]),\nTensor(f32[4, 1, 1]),\n)\n## Return values\nTuple[length=3](\nTensor(f32[4, 1, 1024]),\nTensor(f32[4, 1, 1]),\nTensor(f32[4, 1, 1]),\n)", + "text":"Skipped for fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1, 1024]),\nTensor(f32[4, 1, 1]),\nTensor(f32[4, 1, 1]),\n): Cannot find type promotion rule for op: aten.native_layer_norm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1, 1024]): node.target is not OpOverload. Got type: \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1, 1024])\n## Return values\nTensor(f32[4, 1, 1024])", + "text":"Skipped for fx.Node()[call_function]:Tensor(f32[4, 1, 1024]): node.target is not OpOverload. Got type: " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024])\n## Return values\nTensor(f32[4, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n## Return values\nTensor(f32[1024, 4096])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4, 4096]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4, 4096])\n## Return values\nTensor(f32[4, 4096])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4, 4096]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 4096])\n## Return values\nTensor(f32[4, 1, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for gelu_24. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1, 4096])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'gelu', [0], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument view_251 is not promoted. Already torch.float32.\nArgument tanh is not promoted. Not mentioned by type promotion rule.\n## Return values\nTensor(f32[4, 1, 4096])", + "text":"Type promotion not needed for gelu_24. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1, 4096]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1, 4096])\n## Return values\nTensor(f32[4, 1, 4096])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1, 4096]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 4096]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 4096])\n## Return values\nTensor(f32[4, 4096])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 4096]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024])\n## Return values\nTensor(f32[4096, 1024])", + "text":"Skipped for fx.Node(aten.t.default)[call_function]:Tensor(f32[4096, 1024]): Cannot find type promotion rule for op: aten.t.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.addmm.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4, 1024])\n## Return values\nTensor(f32[4, 1024])", + "text":"Skipped for fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.addmm.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.view.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 1024])\n## Return values\nTensor(f32[4, 1, 1024])", + "text":"Skipped for fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.view.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1, 1024])\n## Return values\nTensor(f32[4, 1, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Type promotion not needed for add_50. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1, 1024])\nFound type promotion rule: ElementwiseTypePromotionRule('aten', 'add', [0, 1], [], ELEMENTWISE_TYPE_PROMOTION_KIND.DEFAULT)\nArgument clone_73 is not promoted. Already torch.float32.\nArgument clone_75 is not promoted. Already torch.float32.\n## Return values\nTensor(f32[4, 1, 1024])", + "text":"Type promotion not needed for add_50. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.slice.Tensor)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.slice.Tensor\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.slice.Tensor)[call_function]:Tensor(f32[4, 1, 1024])\n## Return values\nTensor(f32[4, 1, 1024])", + "text":"Skipped for fx.Node(aten.slice.Tensor)[call_function]:Tensor(f32[4, 1, 1024]): Cannot find type promotion rule for op: aten.slice.Tensor" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.select.int)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.select.int\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.select.int)[call_function]:Tensor(f32[4, 1024])\n## Return values\nTensor(f32[4, 1024])", + "text":"Skipped for fx.Node(aten.select.int)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.select.int" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.clone.default\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024])\n## Return values\nTensor(f32[4, 1024])", + "text":"Skipped for fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024]): Cannot find type promotion rule for op: aten.clone.default" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Skipped output: not a call_function.\n\n## Additional Message:\n\n## Function Signature\n### Function Signature _TypePromotionInterpreter.run_node\n- self: \n- node: fx.Node(output)[output]:None\n## Return values\nTuple[length=1](\nTensor(f32[4, 1024]),\n)", + "text":"Skipped output: not a call_function." + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"_TypePromotionInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/passes/type_promotion.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":1625 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0015", + "stacks":[] + }, + { + "message":{ + "markdown":"Running InsertTypePromotion pass. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature Transform.run\n- self: \nFor detailed logging of graph modifications by this pass, either set `DiagnosticOptions.verbosity_level` to `logging.DEBUG` or use the environment variable `TORCH_LOGS='onnx_diagnostics'`.\n## Return values\ntorch.fx.GraphModule(GraphModule)", + "text":"Running InsertTypePromotion pass. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"Transform.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/_pass.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":243 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0010", + "stacks":[] + }, + { + "message":{ + "markdown":"Running Modularize pass. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature Transform.run\n- self: \nFor detailed logging of graph modifications by this pass, either set `DiagnosticOptions.verbosity_level` to `logging.DEBUG` or use the environment variable `TORCH_LOGS='onnx_diagnostics'`.\n## Return values\ntorch.fx.GraphModule()", + "text":"Running Modularize pass. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"Transform.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/_pass.py" + }, + "region":{ + "snippet":{ + "text":"@diagnostics.diagnose_call(" + }, + "startLine":243 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0010", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_pos_embed[name=p_trunk_pos_embed]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_pos_embed)[placeholder]:Tensor(f32[1, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_pos_embed[name=p_trunk_pos_embed]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_latent[name=p_trunk_attn_pool_latent]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_latent)[placeholder]:Tensor(f32[1, 1, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_latent[name=p_trunk_attn_pool_latent]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_patch_embed_proj_weight[name=p_trunk_patch_embed_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_patch_embed_proj_weight)[placeholder]:Tensor(f32[1024, 3, 16, 16])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_patch_embed_proj_weight[name=p_trunk_patch_embed_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_patch_embed_proj_bias[name=p_trunk_patch_embed_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_patch_embed_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_patch_embed_proj_bias[name=p_trunk_patch_embed_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_weight[name=p_getattr_l__self___trunk_blocks___0___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_weight[name=p_getattr_l__self___trunk_blocks___0___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_bias[name=p_getattr_l__self___trunk_blocks___0___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=5](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_bias[name=p_getattr_l__self___trunk_blocks___0___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=6](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=7](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___0___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=8](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___0___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___0___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=9](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___0___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_weight[name=p_getattr_l__self___trunk_blocks___0___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=10](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_weight[name=p_getattr_l__self___trunk_blocks___0___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_bias[name=p_getattr_l__self___trunk_blocks___0___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=11](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_bias[name=p_getattr_l__self___trunk_blocks___0___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=12](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=13](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=14](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=15](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm1_weight[name=p_getattr_l__self___trunk_blocks___1___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=16](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm1_weight[name=p_getattr_l__self___trunk_blocks___1___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm1_bias[name=p_getattr_l__self___trunk_blocks___1___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=17](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm1_bias[name=p_getattr_l__self___trunk_blocks___1___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___1___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=18](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___1___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___1___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=19](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___1___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___1___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=20](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___1___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___1___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=21](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___1___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm2_weight[name=p_getattr_l__self___trunk_blocks___1___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=22](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm2_weight[name=p_getattr_l__self___trunk_blocks___1___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm2_bias[name=p_getattr_l__self___trunk_blocks___1___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=23](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm2_bias[name=p_getattr_l__self___trunk_blocks___1___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=24](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=25](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=26](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=27](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm1_weight[name=p_getattr_l__self___trunk_blocks___2___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=28](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm1_weight[name=p_getattr_l__self___trunk_blocks___2___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm1_bias[name=p_getattr_l__self___trunk_blocks___2___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=29](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm1_bias[name=p_getattr_l__self___trunk_blocks___2___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___2___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=30](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___2___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___2___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=31](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___2___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___2___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=32](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___2___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___2___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=33](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___2___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm2_weight[name=p_getattr_l__self___trunk_blocks___2___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=34](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm2_weight[name=p_getattr_l__self___trunk_blocks___2___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm2_bias[name=p_getattr_l__self___trunk_blocks___2___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=35](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm2_bias[name=p_getattr_l__self___trunk_blocks___2___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=36](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=37](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=38](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=39](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm1_weight[name=p_getattr_l__self___trunk_blocks___3___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=40](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm1_weight[name=p_getattr_l__self___trunk_blocks___3___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm1_bias[name=p_getattr_l__self___trunk_blocks___3___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=41](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm1_bias[name=p_getattr_l__self___trunk_blocks___3___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___3___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=42](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___3___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___3___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=43](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___3___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___3___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=44](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___3___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___3___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=45](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___3___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm2_weight[name=p_getattr_l__self___trunk_blocks___3___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=46](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm2_weight[name=p_getattr_l__self___trunk_blocks___3___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm2_bias[name=p_getattr_l__self___trunk_blocks___3___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=47](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm2_bias[name=p_getattr_l__self___trunk_blocks___3___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=48](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=49](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=50](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=51](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm1_weight[name=p_getattr_l__self___trunk_blocks___4___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=52](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm1_weight[name=p_getattr_l__self___trunk_blocks___4___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm1_bias[name=p_getattr_l__self___trunk_blocks___4___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=53](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm1_bias[name=p_getattr_l__self___trunk_blocks___4___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___4___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=54](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___4___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___4___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=55](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___4___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___4___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=56](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___4___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___4___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=57](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___4___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm2_weight[name=p_getattr_l__self___trunk_blocks___4___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=58](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm2_weight[name=p_getattr_l__self___trunk_blocks___4___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm2_bias[name=p_getattr_l__self___trunk_blocks___4___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=59](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm2_bias[name=p_getattr_l__self___trunk_blocks___4___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=60](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=61](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=62](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=63](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm1_weight[name=p_getattr_l__self___trunk_blocks___5___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=64](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm1_weight[name=p_getattr_l__self___trunk_blocks___5___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm1_bias[name=p_getattr_l__self___trunk_blocks___5___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=65](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm1_bias[name=p_getattr_l__self___trunk_blocks___5___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___5___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=66](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___5___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___5___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=67](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___5___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___5___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=68](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___5___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___5___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=69](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___5___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm2_weight[name=p_getattr_l__self___trunk_blocks___5___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=70](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm2_weight[name=p_getattr_l__self___trunk_blocks___5___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm2_bias[name=p_getattr_l__self___trunk_blocks___5___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=71](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm2_bias[name=p_getattr_l__self___trunk_blocks___5___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=72](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=73](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=74](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=75](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm1_weight[name=p_getattr_l__self___trunk_blocks___6___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=76](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm1_weight[name=p_getattr_l__self___trunk_blocks___6___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm1_bias[name=p_getattr_l__self___trunk_blocks___6___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=77](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm1_bias[name=p_getattr_l__self___trunk_blocks___6___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___6___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=78](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___6___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___6___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=79](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___6___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___6___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=80](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___6___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___6___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=81](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___6___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm2_weight[name=p_getattr_l__self___trunk_blocks___6___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=82](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm2_weight[name=p_getattr_l__self___trunk_blocks___6___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm2_bias[name=p_getattr_l__self___trunk_blocks___6___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=83](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm2_bias[name=p_getattr_l__self___trunk_blocks___6___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=84](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=85](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=86](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=87](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm1_weight[name=p_getattr_l__self___trunk_blocks___7___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=88](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm1_weight[name=p_getattr_l__self___trunk_blocks___7___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm1_bias[name=p_getattr_l__self___trunk_blocks___7___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=89](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm1_bias[name=p_getattr_l__self___trunk_blocks___7___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___7___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=90](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___7___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___7___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=91](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___7___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___7___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=92](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___7___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___7___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=93](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___7___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm2_weight[name=p_getattr_l__self___trunk_blocks___7___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=94](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm2_weight[name=p_getattr_l__self___trunk_blocks___7___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm2_bias[name=p_getattr_l__self___trunk_blocks___7___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=95](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm2_bias[name=p_getattr_l__self___trunk_blocks___7___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=96](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=97](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=98](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=99](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm1_weight[name=p_getattr_l__self___trunk_blocks___8___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=100](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm1_weight[name=p_getattr_l__self___trunk_blocks___8___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm1_bias[name=p_getattr_l__self___trunk_blocks___8___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=101](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm1_bias[name=p_getattr_l__self___trunk_blocks___8___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___8___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=102](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___8___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___8___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=103](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___8___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___8___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=104](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___8___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___8___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=105](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___8___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm2_weight[name=p_getattr_l__self___trunk_blocks___8___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=106](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm2_weight[name=p_getattr_l__self___trunk_blocks___8___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm2_bias[name=p_getattr_l__self___trunk_blocks___8___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=107](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm2_bias[name=p_getattr_l__self___trunk_blocks___8___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=108](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=109](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=110](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=111](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm1_weight[name=p_getattr_l__self___trunk_blocks___9___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=112](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm1_weight[name=p_getattr_l__self___trunk_blocks___9___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm1_bias[name=p_getattr_l__self___trunk_blocks___9___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=113](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm1_bias[name=p_getattr_l__self___trunk_blocks___9___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___9___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=114](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___9___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___9___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=115](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___9___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___9___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=116](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___9___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___9___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=117](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___9___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm2_weight[name=p_getattr_l__self___trunk_blocks___9___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=118](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm2_weight[name=p_getattr_l__self___trunk_blocks___9___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm2_bias[name=p_getattr_l__self___trunk_blocks___9___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=119](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm2_bias[name=p_getattr_l__self___trunk_blocks___9___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=120](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=121](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=122](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=123](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm1_weight[name=p_getattr_l__self___trunk_blocks___10___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=124](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm1_weight[name=p_getattr_l__self___trunk_blocks___10___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm1_bias[name=p_getattr_l__self___trunk_blocks___10___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=125](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm1_bias[name=p_getattr_l__self___trunk_blocks___10___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___10___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=126](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___10___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___10___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=127](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___10___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___10___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=128](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___10___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___10___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=129](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___10___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm2_weight[name=p_getattr_l__self___trunk_blocks___10___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=130](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm2_weight[name=p_getattr_l__self___trunk_blocks___10___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm2_bias[name=p_getattr_l__self___trunk_blocks___10___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=131](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm2_bias[name=p_getattr_l__self___trunk_blocks___10___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=132](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=133](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=134](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=135](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm1_weight[name=p_getattr_l__self___trunk_blocks___11___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=136](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm1_weight[name=p_getattr_l__self___trunk_blocks___11___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm1_bias[name=p_getattr_l__self___trunk_blocks___11___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=137](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm1_bias[name=p_getattr_l__self___trunk_blocks___11___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___11___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=138](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___11___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___11___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=139](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___11___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___11___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=140](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___11___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___11___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=141](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___11___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm2_weight[name=p_getattr_l__self___trunk_blocks___11___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=142](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm2_weight[name=p_getattr_l__self___trunk_blocks___11___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm2_bias[name=p_getattr_l__self___trunk_blocks___11___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=143](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm2_bias[name=p_getattr_l__self___trunk_blocks___11___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=144](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=145](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=146](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=147](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm1_weight[name=p_getattr_l__self___trunk_blocks___12___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=148](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm1_weight[name=p_getattr_l__self___trunk_blocks___12___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm1_bias[name=p_getattr_l__self___trunk_blocks___12___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=149](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm1_bias[name=p_getattr_l__self___trunk_blocks___12___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___12___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=150](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___12___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___12___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=151](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___12___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___12___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=152](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___12___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___12___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=153](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___12___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm2_weight[name=p_getattr_l__self___trunk_blocks___12___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=154](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm2_weight[name=p_getattr_l__self___trunk_blocks___12___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm2_bias[name=p_getattr_l__self___trunk_blocks___12___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=155](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm2_bias[name=p_getattr_l__self___trunk_blocks___12___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=156](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=157](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=158](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=159](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm1_weight[name=p_getattr_l__self___trunk_blocks___13___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=160](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm1_weight[name=p_getattr_l__self___trunk_blocks___13___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm1_bias[name=p_getattr_l__self___trunk_blocks___13___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=161](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm1_bias[name=p_getattr_l__self___trunk_blocks___13___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___13___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=162](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___13___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___13___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=163](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___13___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___13___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=164](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___13___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___13___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=165](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___13___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm2_weight[name=p_getattr_l__self___trunk_blocks___13___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=166](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm2_weight[name=p_getattr_l__self___trunk_blocks___13___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm2_bias[name=p_getattr_l__self___trunk_blocks___13___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=167](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm2_bias[name=p_getattr_l__self___trunk_blocks___13___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=168](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=169](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=170](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=171](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm1_weight[name=p_getattr_l__self___trunk_blocks___14___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=172](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm1_weight[name=p_getattr_l__self___trunk_blocks___14___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm1_bias[name=p_getattr_l__self___trunk_blocks___14___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=173](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm1_bias[name=p_getattr_l__self___trunk_blocks___14___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___14___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=174](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___14___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___14___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=175](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___14___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___14___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=176](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___14___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___14___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=177](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___14___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm2_weight[name=p_getattr_l__self___trunk_blocks___14___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=178](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm2_weight[name=p_getattr_l__self___trunk_blocks___14___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm2_bias[name=p_getattr_l__self___trunk_blocks___14___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=179](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm2_bias[name=p_getattr_l__self___trunk_blocks___14___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=180](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=181](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=182](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=183](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm1_weight[name=p_getattr_l__self___trunk_blocks___15___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=184](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm1_weight[name=p_getattr_l__self___trunk_blocks___15___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm1_bias[name=p_getattr_l__self___trunk_blocks___15___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=185](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm1_bias[name=p_getattr_l__self___trunk_blocks___15___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___15___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=186](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___15___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___15___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=187](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___15___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___15___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=188](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___15___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___15___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=189](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___15___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm2_weight[name=p_getattr_l__self___trunk_blocks___15___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=190](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm2_weight[name=p_getattr_l__self___trunk_blocks___15___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm2_bias[name=p_getattr_l__self___trunk_blocks___15___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=191](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm2_bias[name=p_getattr_l__self___trunk_blocks___15___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=192](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=193](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=194](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=195](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm1_weight[name=p_getattr_l__self___trunk_blocks___16___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=196](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm1_weight[name=p_getattr_l__self___trunk_blocks___16___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm1_bias[name=p_getattr_l__self___trunk_blocks___16___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=197](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm1_bias[name=p_getattr_l__self___trunk_blocks___16___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___16___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=198](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___16___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___16___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=199](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___16___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___16___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=200](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___16___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___16___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=201](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___16___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm2_weight[name=p_getattr_l__self___trunk_blocks___16___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=202](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm2_weight[name=p_getattr_l__self___trunk_blocks___16___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm2_bias[name=p_getattr_l__self___trunk_blocks___16___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=203](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm2_bias[name=p_getattr_l__self___trunk_blocks___16___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=204](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=205](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=206](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=207](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm1_weight[name=p_getattr_l__self___trunk_blocks___17___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=208](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm1_weight[name=p_getattr_l__self___trunk_blocks___17___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm1_bias[name=p_getattr_l__self___trunk_blocks___17___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=209](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm1_bias[name=p_getattr_l__self___trunk_blocks___17___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___17___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=210](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___17___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___17___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=211](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___17___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___17___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=212](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___17___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___17___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=213](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___17___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm2_weight[name=p_getattr_l__self___trunk_blocks___17___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=214](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm2_weight[name=p_getattr_l__self___trunk_blocks___17___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm2_bias[name=p_getattr_l__self___trunk_blocks___17___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=215](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm2_bias[name=p_getattr_l__self___trunk_blocks___17___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=216](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=217](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=218](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=219](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm1_weight[name=p_getattr_l__self___trunk_blocks___18___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=220](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm1_weight[name=p_getattr_l__self___trunk_blocks___18___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm1_bias[name=p_getattr_l__self___trunk_blocks___18___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=221](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm1_bias[name=p_getattr_l__self___trunk_blocks___18___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___18___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=222](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___18___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___18___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=223](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___18___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___18___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=224](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___18___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___18___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=225](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___18___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm2_weight[name=p_getattr_l__self___trunk_blocks___18___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=226](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm2_weight[name=p_getattr_l__self___trunk_blocks___18___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm2_bias[name=p_getattr_l__self___trunk_blocks___18___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=227](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm2_bias[name=p_getattr_l__self___trunk_blocks___18___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=228](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=229](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=230](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=231](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm1_weight[name=p_getattr_l__self___trunk_blocks___19___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=232](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm1_weight[name=p_getattr_l__self___trunk_blocks___19___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm1_bias[name=p_getattr_l__self___trunk_blocks___19___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=233](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm1_bias[name=p_getattr_l__self___trunk_blocks___19___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___19___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=234](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___19___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___19___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=235](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___19___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___19___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=236](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___19___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___19___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=237](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___19___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm2_weight[name=p_getattr_l__self___trunk_blocks___19___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=238](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm2_weight[name=p_getattr_l__self___trunk_blocks___19___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm2_bias[name=p_getattr_l__self___trunk_blocks___19___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=239](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm2_bias[name=p_getattr_l__self___trunk_blocks___19___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=240](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=241](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=242](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=243](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm1_weight[name=p_getattr_l__self___trunk_blocks___20___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=244](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm1_weight[name=p_getattr_l__self___trunk_blocks___20___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm1_bias[name=p_getattr_l__self___trunk_blocks___20___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=245](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm1_bias[name=p_getattr_l__self___trunk_blocks___20___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___20___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=246](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___20___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___20___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=247](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___20___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___20___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=248](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___20___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___20___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=249](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___20___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm2_weight[name=p_getattr_l__self___trunk_blocks___20___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=250](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm2_weight[name=p_getattr_l__self___trunk_blocks___20___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm2_bias[name=p_getattr_l__self___trunk_blocks___20___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=251](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm2_bias[name=p_getattr_l__self___trunk_blocks___20___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=252](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=253](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=254](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=255](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm1_weight[name=p_getattr_l__self___trunk_blocks___21___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=256](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm1_weight[name=p_getattr_l__self___trunk_blocks___21___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm1_bias[name=p_getattr_l__self___trunk_blocks___21___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=257](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm1_bias[name=p_getattr_l__self___trunk_blocks___21___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___21___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=258](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___21___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___21___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=259](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___21___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___21___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=260](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___21___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___21___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=261](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___21___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm2_weight[name=p_getattr_l__self___trunk_blocks___21___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=262](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm2_weight[name=p_getattr_l__self___trunk_blocks___21___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm2_bias[name=p_getattr_l__self___trunk_blocks___21___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=263](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm2_bias[name=p_getattr_l__self___trunk_blocks___21___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=264](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=265](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=266](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=267](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm1_weight[name=p_getattr_l__self___trunk_blocks___22___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=268](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm1_weight[name=p_getattr_l__self___trunk_blocks___22___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm1_bias[name=p_getattr_l__self___trunk_blocks___22___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=269](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm1_bias[name=p_getattr_l__self___trunk_blocks___22___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___22___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=270](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___22___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___22___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=271](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___22___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___22___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=272](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___22___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___22___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=273](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___22___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm2_weight[name=p_getattr_l__self___trunk_blocks___22___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=274](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm2_weight[name=p_getattr_l__self___trunk_blocks___22___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm2_bias[name=p_getattr_l__self___trunk_blocks___22___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=275](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm2_bias[name=p_getattr_l__self___trunk_blocks___22___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=276](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=277](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=278](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=279](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm1_weight[name=p_getattr_l__self___trunk_blocks___23___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=280](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm1_weight[name=p_getattr_l__self___trunk_blocks___23___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm1_bias[name=p_getattr_l__self___trunk_blocks___23___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=281](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm1_bias[name=p_getattr_l__self___trunk_blocks___23___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___23___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=282](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___23___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___23___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=283](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___23___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___23___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=284](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___23___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___23___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=285](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___23___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm2_weight[name=p_getattr_l__self___trunk_blocks___23___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=286](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm2_weight[name=p_getattr_l__self___trunk_blocks___23___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm2_bias[name=p_getattr_l__self___trunk_blocks___23___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=287](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm2_bias[name=p_getattr_l__self___trunk_blocks___23___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=288](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=289](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=290](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=291](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_norm_weight[name=p_trunk_norm_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_norm_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=292](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_norm_weight[name=p_trunk_norm_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_norm_bias[name=p_trunk_norm_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_norm_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=293](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_norm_bias[name=p_trunk_norm_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_q_weight[name=p_trunk_attn_pool_q_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_q_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=294](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_q_weight[name=p_trunk_attn_pool_q_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_q_bias[name=p_trunk_attn_pool_q_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_q_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=295](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_q_bias[name=p_trunk_attn_pool_q_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_kv_weight[name=p_trunk_attn_pool_kv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_kv_weight)[placeholder]:Tensor(f32[2048, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=296](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_kv_weight[name=p_trunk_attn_pool_kv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_kv_bias[name=p_trunk_attn_pool_kv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_kv_bias)[placeholder]:Tensor(f32[2048])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=297](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_kv_bias[name=p_trunk_attn_pool_kv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_proj_weight[name=p_trunk_attn_pool_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=298](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_proj_weight[name=p_trunk_attn_pool_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_proj_bias[name=p_trunk_attn_pool_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=299](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_proj_bias[name=p_trunk_attn_pool_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_norm_weight[name=p_trunk_attn_pool_norm_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_norm_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=300](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_norm_weight[name=p_trunk_attn_pool_norm_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_norm_bias[name=p_trunk_attn_pool_norm_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_norm_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=301](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_norm_bias[name=p_trunk_attn_pool_norm_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc1_weight[name=p_trunk_attn_pool_mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=302](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc1_weight[name=p_trunk_attn_pool_mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc1_bias[name=p_trunk_attn_pool_mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=303](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc1_bias[name=p_trunk_attn_pool_mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc2_weight[name=p_trunk_attn_pool_mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=304](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc2_weight[name=p_trunk_attn_pool_mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc2_bias[name=p_trunk_attn_pool_mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=305](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc2_bias[name=p_trunk_attn_pool_mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:x[name=x]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(x)[placeholder]:Tensor(f32[4, 3, 512, 512])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=306](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:x[name=x]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:x[name=x]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(x)[placeholder]:Tensor(f32[4, 3, 512, 512])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## Return values\n", + "text":"FX Node: placeholder:x[name=x]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_patch_embed_proj_weight[name=p_trunk_patch_embed_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_patch_embed_proj_weight)[placeholder]:Tensor(f32[1024, 3, 16, 16])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_patch_embed_proj_weight[name=p_trunk_patch_embed_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_patch_embed_proj_bias[name=p_trunk_patch_embed_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_patch_embed_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_patch_embed_proj_bias[name=p_trunk_patch_embed_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_pos_embed[name=p_trunk_pos_embed]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_pos_embed)[placeholder]:Tensor(f32[1, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_pos_embed[name=p_trunk_pos_embed]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_weight[name=p_getattr_l__self___trunk_blocks___0___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_weight[name=p_getattr_l__self___trunk_blocks___0___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_bias[name=p_getattr_l__self___trunk_blocks___0___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=5](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_bias[name=p_getattr_l__self___trunk_blocks___0___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=6](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=7](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___0___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=8](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___0___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___0___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=9](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___0___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_weight[name=p_getattr_l__self___trunk_blocks___0___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=10](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_weight[name=p_getattr_l__self___trunk_blocks___0___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_bias[name=p_getattr_l__self___trunk_blocks___0___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=11](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_bias[name=p_getattr_l__self___trunk_blocks___0___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=12](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=13](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=14](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=15](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm1_weight[name=p_getattr_l__self___trunk_blocks___1___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=16](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm1_weight[name=p_getattr_l__self___trunk_blocks___1___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm1_bias[name=p_getattr_l__self___trunk_blocks___1___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=17](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm1_bias[name=p_getattr_l__self___trunk_blocks___1___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___1___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=18](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___1___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___1___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=19](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___1___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___1___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=20](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___1___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___1___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=21](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___1___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm2_weight[name=p_getattr_l__self___trunk_blocks___1___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=22](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm2_weight[name=p_getattr_l__self___trunk_blocks___1___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm2_bias[name=p_getattr_l__self___trunk_blocks___1___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=23](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm2_bias[name=p_getattr_l__self___trunk_blocks___1___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=24](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=25](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=26](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=27](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm1_weight[name=p_getattr_l__self___trunk_blocks___2___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=28](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm1_weight[name=p_getattr_l__self___trunk_blocks___2___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm1_bias[name=p_getattr_l__self___trunk_blocks___2___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=29](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm1_bias[name=p_getattr_l__self___trunk_blocks___2___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___2___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=30](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___2___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___2___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=31](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___2___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___2___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=32](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___2___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___2___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=33](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___2___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm2_weight[name=p_getattr_l__self___trunk_blocks___2___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=34](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm2_weight[name=p_getattr_l__self___trunk_blocks___2___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm2_bias[name=p_getattr_l__self___trunk_blocks___2___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=35](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm2_bias[name=p_getattr_l__self___trunk_blocks___2___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=36](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=37](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=38](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=39](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm1_weight[name=p_getattr_l__self___trunk_blocks___3___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=40](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm1_weight[name=p_getattr_l__self___trunk_blocks___3___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm1_bias[name=p_getattr_l__self___trunk_blocks___3___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=41](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm1_bias[name=p_getattr_l__self___trunk_blocks___3___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___3___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=42](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___3___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___3___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=43](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___3___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___3___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=44](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___3___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___3___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=45](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___3___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm2_weight[name=p_getattr_l__self___trunk_blocks___3___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=46](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm2_weight[name=p_getattr_l__self___trunk_blocks___3___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm2_bias[name=p_getattr_l__self___trunk_blocks___3___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=47](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm2_bias[name=p_getattr_l__self___trunk_blocks___3___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=48](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=49](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=50](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=51](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm1_weight[name=p_getattr_l__self___trunk_blocks___4___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=52](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm1_weight[name=p_getattr_l__self___trunk_blocks___4___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm1_bias[name=p_getattr_l__self___trunk_blocks___4___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=53](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm1_bias[name=p_getattr_l__self___trunk_blocks___4___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___4___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=54](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___4___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___4___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=55](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___4___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___4___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=56](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___4___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___4___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=57](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___4___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm2_weight[name=p_getattr_l__self___trunk_blocks___4___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=58](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm2_weight[name=p_getattr_l__self___trunk_blocks___4___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm2_bias[name=p_getattr_l__self___trunk_blocks___4___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=59](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm2_bias[name=p_getattr_l__self___trunk_blocks___4___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=60](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=61](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=62](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=63](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm1_weight[name=p_getattr_l__self___trunk_blocks___5___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=64](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm1_weight[name=p_getattr_l__self___trunk_blocks___5___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm1_bias[name=p_getattr_l__self___trunk_blocks___5___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=65](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm1_bias[name=p_getattr_l__self___trunk_blocks___5___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___5___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=66](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___5___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___5___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=67](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___5___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___5___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=68](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___5___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___5___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=69](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___5___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm2_weight[name=p_getattr_l__self___trunk_blocks___5___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=70](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm2_weight[name=p_getattr_l__self___trunk_blocks___5___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm2_bias[name=p_getattr_l__self___trunk_blocks___5___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=71](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm2_bias[name=p_getattr_l__self___trunk_blocks___5___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=72](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=73](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=74](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=75](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm1_weight[name=p_getattr_l__self___trunk_blocks___6___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=76](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm1_weight[name=p_getattr_l__self___trunk_blocks___6___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm1_bias[name=p_getattr_l__self___trunk_blocks___6___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=77](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm1_bias[name=p_getattr_l__self___trunk_blocks___6___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___6___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=78](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___6___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___6___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=79](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___6___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___6___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=80](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___6___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___6___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=81](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___6___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm2_weight[name=p_getattr_l__self___trunk_blocks___6___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=82](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm2_weight[name=p_getattr_l__self___trunk_blocks___6___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm2_bias[name=p_getattr_l__self___trunk_blocks___6___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=83](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm2_bias[name=p_getattr_l__self___trunk_blocks___6___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=84](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=85](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=86](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=87](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm1_weight[name=p_getattr_l__self___trunk_blocks___7___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=88](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm1_weight[name=p_getattr_l__self___trunk_blocks___7___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm1_bias[name=p_getattr_l__self___trunk_blocks___7___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=89](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm1_bias[name=p_getattr_l__self___trunk_blocks___7___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___7___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=90](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___7___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___7___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=91](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___7___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___7___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=92](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___7___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___7___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=93](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___7___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm2_weight[name=p_getattr_l__self___trunk_blocks___7___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=94](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm2_weight[name=p_getattr_l__self___trunk_blocks___7___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm2_bias[name=p_getattr_l__self___trunk_blocks___7___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=95](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm2_bias[name=p_getattr_l__self___trunk_blocks___7___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=96](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=97](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=98](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=99](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm1_weight[name=p_getattr_l__self___trunk_blocks___8___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=100](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm1_weight[name=p_getattr_l__self___trunk_blocks___8___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm1_bias[name=p_getattr_l__self___trunk_blocks___8___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=101](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm1_bias[name=p_getattr_l__self___trunk_blocks___8___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___8___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=102](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___8___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___8___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=103](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___8___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___8___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=104](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___8___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___8___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=105](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___8___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm2_weight[name=p_getattr_l__self___trunk_blocks___8___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=106](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm2_weight[name=p_getattr_l__self___trunk_blocks___8___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm2_bias[name=p_getattr_l__self___trunk_blocks___8___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=107](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm2_bias[name=p_getattr_l__self___trunk_blocks___8___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=108](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=109](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=110](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=111](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm1_weight[name=p_getattr_l__self___trunk_blocks___9___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=112](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm1_weight[name=p_getattr_l__self___trunk_blocks___9___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm1_bias[name=p_getattr_l__self___trunk_blocks___9___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=113](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm1_bias[name=p_getattr_l__self___trunk_blocks___9___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___9___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=114](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___9___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___9___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=115](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___9___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___9___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=116](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___9___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___9___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=117](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___9___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm2_weight[name=p_getattr_l__self___trunk_blocks___9___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=118](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm2_weight[name=p_getattr_l__self___trunk_blocks___9___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm2_bias[name=p_getattr_l__self___trunk_blocks___9___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=119](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm2_bias[name=p_getattr_l__self___trunk_blocks___9___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=120](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=121](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=122](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=123](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm1_weight[name=p_getattr_l__self___trunk_blocks___10___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=124](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm1_weight[name=p_getattr_l__self___trunk_blocks___10___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm1_bias[name=p_getattr_l__self___trunk_blocks___10___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=125](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm1_bias[name=p_getattr_l__self___trunk_blocks___10___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___10___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=126](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___10___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___10___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=127](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___10___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___10___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=128](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___10___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___10___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=129](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___10___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm2_weight[name=p_getattr_l__self___trunk_blocks___10___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=130](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm2_weight[name=p_getattr_l__self___trunk_blocks___10___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm2_bias[name=p_getattr_l__self___trunk_blocks___10___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=131](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm2_bias[name=p_getattr_l__self___trunk_blocks___10___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=132](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=133](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=134](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=135](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm1_weight[name=p_getattr_l__self___trunk_blocks___11___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=136](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm1_weight[name=p_getattr_l__self___trunk_blocks___11___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm1_bias[name=p_getattr_l__self___trunk_blocks___11___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=137](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm1_bias[name=p_getattr_l__self___trunk_blocks___11___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___11___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=138](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___11___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___11___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=139](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___11___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___11___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=140](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___11___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___11___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=141](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___11___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm2_weight[name=p_getattr_l__self___trunk_blocks___11___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=142](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm2_weight[name=p_getattr_l__self___trunk_blocks___11___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm2_bias[name=p_getattr_l__self___trunk_blocks___11___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=143](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm2_bias[name=p_getattr_l__self___trunk_blocks___11___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=144](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=145](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=146](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=147](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm1_weight[name=p_getattr_l__self___trunk_blocks___12___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=148](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm1_weight[name=p_getattr_l__self___trunk_blocks___12___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm1_bias[name=p_getattr_l__self___trunk_blocks___12___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=149](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm1_bias[name=p_getattr_l__self___trunk_blocks___12___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___12___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=150](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___12___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___12___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=151](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___12___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___12___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=152](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___12___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___12___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=153](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___12___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm2_weight[name=p_getattr_l__self___trunk_blocks___12___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=154](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm2_weight[name=p_getattr_l__self___trunk_blocks___12___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm2_bias[name=p_getattr_l__self___trunk_blocks___12___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=155](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm2_bias[name=p_getattr_l__self___trunk_blocks___12___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=156](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=157](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=158](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=159](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm1_weight[name=p_getattr_l__self___trunk_blocks___13___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=160](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm1_weight[name=p_getattr_l__self___trunk_blocks___13___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm1_bias[name=p_getattr_l__self___trunk_blocks___13___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=161](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm1_bias[name=p_getattr_l__self___trunk_blocks___13___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___13___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=162](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___13___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___13___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=163](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___13___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___13___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=164](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___13___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___13___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=165](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___13___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm2_weight[name=p_getattr_l__self___trunk_blocks___13___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=166](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm2_weight[name=p_getattr_l__self___trunk_blocks___13___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm2_bias[name=p_getattr_l__self___trunk_blocks___13___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=167](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm2_bias[name=p_getattr_l__self___trunk_blocks___13___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=168](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=169](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=170](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=171](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm1_weight[name=p_getattr_l__self___trunk_blocks___14___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=172](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm1_weight[name=p_getattr_l__self___trunk_blocks___14___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm1_bias[name=p_getattr_l__self___trunk_blocks___14___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=173](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm1_bias[name=p_getattr_l__self___trunk_blocks___14___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___14___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=174](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___14___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___14___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=175](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___14___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___14___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=176](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___14___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___14___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=177](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___14___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm2_weight[name=p_getattr_l__self___trunk_blocks___14___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=178](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm2_weight[name=p_getattr_l__self___trunk_blocks___14___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm2_bias[name=p_getattr_l__self___trunk_blocks___14___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=179](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm2_bias[name=p_getattr_l__self___trunk_blocks___14___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=180](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=181](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=182](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=183](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm1_weight[name=p_getattr_l__self___trunk_blocks___15___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=184](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm1_weight[name=p_getattr_l__self___trunk_blocks___15___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm1_bias[name=p_getattr_l__self___trunk_blocks___15___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=185](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm1_bias[name=p_getattr_l__self___trunk_blocks___15___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___15___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=186](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___15___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___15___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=187](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___15___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___15___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=188](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___15___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___15___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=189](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___15___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm2_weight[name=p_getattr_l__self___trunk_blocks___15___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=190](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm2_weight[name=p_getattr_l__self___trunk_blocks___15___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm2_bias[name=p_getattr_l__self___trunk_blocks___15___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=191](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm2_bias[name=p_getattr_l__self___trunk_blocks___15___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=192](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=193](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=194](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=195](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm1_weight[name=p_getattr_l__self___trunk_blocks___16___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=196](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm1_weight[name=p_getattr_l__self___trunk_blocks___16___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm1_bias[name=p_getattr_l__self___trunk_blocks___16___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=197](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm1_bias[name=p_getattr_l__self___trunk_blocks___16___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___16___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=198](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___16___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___16___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=199](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___16___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___16___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=200](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___16___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___16___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=201](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___16___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm2_weight[name=p_getattr_l__self___trunk_blocks___16___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=202](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm2_weight[name=p_getattr_l__self___trunk_blocks___16___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm2_bias[name=p_getattr_l__self___trunk_blocks___16___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=203](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm2_bias[name=p_getattr_l__self___trunk_blocks___16___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=204](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=205](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=206](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=207](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm1_weight[name=p_getattr_l__self___trunk_blocks___17___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=208](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm1_weight[name=p_getattr_l__self___trunk_blocks___17___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm1_bias[name=p_getattr_l__self___trunk_blocks___17___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=209](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm1_bias[name=p_getattr_l__self___trunk_blocks___17___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___17___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=210](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___17___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___17___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=211](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___17___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___17___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=212](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___17___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___17___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=213](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___17___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm2_weight[name=p_getattr_l__self___trunk_blocks___17___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=214](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm2_weight[name=p_getattr_l__self___trunk_blocks___17___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm2_bias[name=p_getattr_l__self___trunk_blocks___17___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=215](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm2_bias[name=p_getattr_l__self___trunk_blocks___17___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=216](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=217](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=218](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=219](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm1_weight[name=p_getattr_l__self___trunk_blocks___18___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=220](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm1_weight[name=p_getattr_l__self___trunk_blocks___18___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm1_bias[name=p_getattr_l__self___trunk_blocks___18___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=221](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm1_bias[name=p_getattr_l__self___trunk_blocks___18___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___18___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=222](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___18___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___18___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=223](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___18___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___18___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=224](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___18___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___18___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=225](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___18___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm2_weight[name=p_getattr_l__self___trunk_blocks___18___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=226](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm2_weight[name=p_getattr_l__self___trunk_blocks___18___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm2_bias[name=p_getattr_l__self___trunk_blocks___18___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=227](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm2_bias[name=p_getattr_l__self___trunk_blocks___18___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=228](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=229](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=230](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=231](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm1_weight[name=p_getattr_l__self___trunk_blocks___19___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=232](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm1_weight[name=p_getattr_l__self___trunk_blocks___19___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm1_bias[name=p_getattr_l__self___trunk_blocks___19___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=233](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm1_bias[name=p_getattr_l__self___trunk_blocks___19___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___19___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=234](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___19___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___19___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=235](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___19___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___19___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=236](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___19___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___19___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=237](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___19___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm2_weight[name=p_getattr_l__self___trunk_blocks___19___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=238](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm2_weight[name=p_getattr_l__self___trunk_blocks___19___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm2_bias[name=p_getattr_l__self___trunk_blocks___19___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=239](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm2_bias[name=p_getattr_l__self___trunk_blocks___19___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=240](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=241](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=242](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=243](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm1_weight[name=p_getattr_l__self___trunk_blocks___20___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=244](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm1_weight[name=p_getattr_l__self___trunk_blocks___20___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm1_bias[name=p_getattr_l__self___trunk_blocks___20___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=245](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm1_bias[name=p_getattr_l__self___trunk_blocks___20___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___20___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=246](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___20___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___20___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=247](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___20___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___20___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=248](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___20___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___20___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=249](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___20___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm2_weight[name=p_getattr_l__self___trunk_blocks___20___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=250](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm2_weight[name=p_getattr_l__self___trunk_blocks___20___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm2_bias[name=p_getattr_l__self___trunk_blocks___20___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=251](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm2_bias[name=p_getattr_l__self___trunk_blocks___20___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=252](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=253](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=254](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=255](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm1_weight[name=p_getattr_l__self___trunk_blocks___21___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=256](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm1_weight[name=p_getattr_l__self___trunk_blocks___21___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm1_bias[name=p_getattr_l__self___trunk_blocks___21___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=257](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm1_bias[name=p_getattr_l__self___trunk_blocks___21___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___21___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=258](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___21___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___21___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=259](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___21___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___21___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=260](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___21___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___21___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=261](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___21___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm2_weight[name=p_getattr_l__self___trunk_blocks___21___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=262](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm2_weight[name=p_getattr_l__self___trunk_blocks___21___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm2_bias[name=p_getattr_l__self___trunk_blocks___21___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=263](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm2_bias[name=p_getattr_l__self___trunk_blocks___21___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=264](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=265](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=266](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=267](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm1_weight[name=p_getattr_l__self___trunk_blocks___22___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=268](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm1_weight[name=p_getattr_l__self___trunk_blocks___22___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm1_bias[name=p_getattr_l__self___trunk_blocks___22___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=269](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm1_bias[name=p_getattr_l__self___trunk_blocks___22___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___22___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=270](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___22___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___22___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=271](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___22___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___22___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=272](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___22___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___22___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=273](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___22___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm2_weight[name=p_getattr_l__self___trunk_blocks___22___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=274](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm2_weight[name=p_getattr_l__self___trunk_blocks___22___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm2_bias[name=p_getattr_l__self___trunk_blocks___22___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=275](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm2_bias[name=p_getattr_l__self___trunk_blocks___22___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=276](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=277](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=278](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=279](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm1_weight[name=p_getattr_l__self___trunk_blocks___23___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=280](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm1_weight[name=p_getattr_l__self___trunk_blocks___23___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm1_bias[name=p_getattr_l__self___trunk_blocks___23___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=281](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm1_bias[name=p_getattr_l__self___trunk_blocks___23___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___23___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=282](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___23___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___23___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=283](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___23___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___23___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=284](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___23___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___23___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=285](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___23___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm2_weight[name=p_getattr_l__self___trunk_blocks___23___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=286](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm2_weight[name=p_getattr_l__self___trunk_blocks___23___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm2_bias[name=p_getattr_l__self___trunk_blocks___23___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=287](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm2_bias[name=p_getattr_l__self___trunk_blocks___23___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=288](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=289](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=290](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=291](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_norm_weight[name=p_trunk_norm_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_norm_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=292](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_norm_weight[name=p_trunk_norm_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_norm_bias[name=p_trunk_norm_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_norm_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=293](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_norm_bias[name=p_trunk_norm_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_latent[name=p_trunk_attn_pool_latent]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_latent)[placeholder]:Tensor(f32[1, 1, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=294](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_latent[name=p_trunk_attn_pool_latent]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_q_weight[name=p_trunk_attn_pool_q_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_q_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=295](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_q_weight[name=p_trunk_attn_pool_q_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_q_bias[name=p_trunk_attn_pool_q_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_q_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=296](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_q_bias[name=p_trunk_attn_pool_q_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_kv_weight[name=p_trunk_attn_pool_kv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_kv_weight)[placeholder]:Tensor(f32[2048, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=297](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_kv_weight[name=p_trunk_attn_pool_kv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_kv_bias[name=p_trunk_attn_pool_kv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_kv_bias)[placeholder]:Tensor(f32[2048])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=298](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_kv_bias[name=p_trunk_attn_pool_kv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_proj_weight[name=p_trunk_attn_pool_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=299](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_proj_weight[name=p_trunk_attn_pool_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_proj_bias[name=p_trunk_attn_pool_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=300](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_proj_bias[name=p_trunk_attn_pool_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_norm_weight[name=p_trunk_attn_pool_norm_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_norm_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=301](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_norm_weight[name=p_trunk_attn_pool_norm_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_norm_bias[name=p_trunk_attn_pool_norm_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_norm_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=302](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_norm_bias[name=p_trunk_attn_pool_norm_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc1_weight[name=p_trunk_attn_pool_mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=303](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc1_weight[name=p_trunk_attn_pool_mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc1_bias[name=p_trunk_attn_pool_mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=304](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc1_bias[name=p_trunk_attn_pool_mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc2_weight[name=p_trunk_attn_pool_mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=305](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc2_weight[name=p_trunk_attn_pool_mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc2_bias[name=p_trunk_attn_pool_mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_attn_pool_mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=306](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_attn_pool_mlp_fc2_bias[name=p_trunk_attn_pool_mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:x[name=x]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(x)[placeholder]:Tensor(f32[4, 3, 512, 512])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_patch_embed_PatchEmbed)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## Return values\n", + "text":"FX Node: placeholder:x[name=x]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_patch_embed_proj_weight[name=p_trunk_patch_embed_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_patch_embed_proj_weight)[placeholder]:Tensor(f32[1024, 3, 16, 16])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_patch_embed_PatchEmbed)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_patch_embed_proj_weight[name=p_trunk_patch_embed_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_patch_embed_proj_bias[name=p_trunk_patch_embed_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_patch_embed_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_patch_embed_PatchEmbed)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_patch_embed_proj_bias[name=p_trunk_patch_embed_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:x[name=x]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(x)[placeholder]:Tensor(f32[4, 3, 512, 512])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_conv_Conv2d)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## Return values\n", + "text":"FX Node: placeholder:x[name=x]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_patch_embed_proj_weight[name=p_trunk_patch_embed_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_patch_embed_proj_weight)[placeholder]:Tensor(f32[1024, 3, 16, 16])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_conv_Conv2d)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_patch_embed_proj_weight[name=p_trunk_patch_embed_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_trunk_patch_embed_proj_bias[name=p_trunk_patch_embed_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_trunk_patch_embed_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_conv_Conv2d)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_trunk_patch_embed_proj_bias[name=p_trunk_patch_embed_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.convolution.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.convolution.default)[call_function]:Tensor(f32[4, 1024, 32, 32])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::convolution.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.convolution.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.convolution.default. \nONNX Node: aten_convolution[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.convolution.default)[call_function]:Tensor(f32[4, 1024, 32, 32])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::convolution.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=9](\n`TorchScriptTensor(f32[4, 3, 512, 512])`,\n`TorchScriptTensor(f32[1024, 3, 16, 16])`,\n`TorchScriptTensor(f32[1024])`,\nList[length=2](\n16,\n16,\n),\nList[length=2](\n0,\n0,\n),\nList[length=2](\n1,\n1,\n),\nFalse,\nList[length=2](\n0,\n0,\n),\n1,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_convolution)`\nmatch score: -1\n## Return values\n`TracedOnnxFunction(aten_convolution)`", + "text":"FX Node: aten.convolution.default. \nONNX Node: aten_convolution[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.convolution.default[name=convolution]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.convolution.default)[call_function]:Tensor(f32[4, 1024, 32, 32])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_conv_Conv2d)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 827, in forward_features\n x = self.patch_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/patch_embed.py\", line 131, in forward\n x = self.proj(x)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.convolution.default[name=convolution]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: output:output[name=output]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(output)[output]:None\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_conv_Conv2d)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\nconvolution: `TorchScriptTensor(f32[4, 1024, 32, 32])`,\n)\n## Return values\n", + "text":"FX Node: output:output[name=output]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: torch_nn_modules_conv_Conv2d. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_conv_Conv2d)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Return values\n", + "text":"FX Graph: torch_nn_modules_conv_Conv2d. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:trunk_patch_embed_proj_1[name=trunk_patch_embed_proj_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(trunk_patch_embed_proj_1)[call_module]:Tensor(f32[4, 1024, 32, 32])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_patch_embed_PatchEmbed)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 827, in forward_features\n x = self.patch_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/patch_embed.py\", line 131, in forward\n x = self.proj(x)\n\n```\n## Return values\n", + "text":"FX Node: call_module:trunk_patch_embed_proj_1[name=trunk_patch_embed_proj_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.view.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.view.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor(f32[4, 1024, 32, 32])`,\nList[length=3](\n4,\n1024,\n1024,\n),\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_view)`\nmatch score: 2\n## Return values\n`TracedOnnxFunction(aten_view)`", + "text":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.view.default[name=view]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_patch_embed_PatchEmbed)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\ntrunk_patch_embed_proj_1: `TorchScriptTensor(f32[4, 1024, 32, 32])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 827, in forward_features\n x = self.patch_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/patch_embed.py\", line 133, in forward\n x = x.flatten(2).transpose(1, 2) # NCHW -> NLC\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.view.default[name=view]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.transpose.int' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 1024])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::transpose.int, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.transpose.int' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.transpose.int. \nONNX Node: aten_transpose[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 1024])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::transpose.int, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=3](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\n1,\n2,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_transpose)`\nmatch score: 1\n## Return values\n`TracedOnnxFunction(aten_transpose)`", + "text":"FX Node: aten.transpose.int. \nONNX Node: aten_transpose[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.transpose.int[name=transpose]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_patch_embed_PatchEmbed)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=5](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\ntrunk_patch_embed_proj_1: `TorchScriptTensor(f32[4, 1024, 32, 32])`,\nview: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 827, in forward_features\n x = self.patch_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/patch_embed.py\", line 133, in forward\n x = x.flatten(2).transpose(1, 2) # NCHW -> NLC\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.transpose.int[name=transpose]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: output:output[name=output]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(output)[output]:None\n- fx_graph_module: torch.fx.GraphModule(timm_layers_patch_embed_PatchEmbed)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=6](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\ntrunk_patch_embed_proj_1: `TorchScriptTensor(f32[4, 1024, 32, 32])`,\nview: `TorchScriptTensor(f32[4, 1024, 1024])`,\ntranspose: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: output:output[name=output]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: timm_layers_patch_embed_PatchEmbed. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(timm_layers_patch_embed_PatchEmbed)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Return values\n", + "text":"FX Graph: timm_layers_patch_embed_PatchEmbed. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:trunk_patch_embed_1[name=trunk_patch_embed_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(trunk_patch_embed_1)[call_module]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=307](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 827, in forward_features\n x = self.patch_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/patch_embed.py\", line 131, in forward\n x = self.proj(x)\n\n```\n## Return values\n", + "text":"FX Node: call_module:trunk_patch_embed_1[name=trunk_patch_embed_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.add.Tensor' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\n- diagnostic_context: \n## Return values\nList[length=2](\nregistration.ONNXFunction(aten::add.Tensor, is_custom=False, is_complex=False),\nregistration.ONNXFunction(aten::add.Tensor, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.add.Tensor' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.add.Tensor. \nONNX Node: aten_add[opset=pkg.onnxscript.torch_lib;is_custom=False]. \nONNX Node: aten_logical_or[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\n- default_and_custom_functions: List[length=2](\nregistration.ONNXFunction(aten::add.Tensor, is_custom=False, is_complex=False),\nregistration.ONNXFunction(aten::add.Tensor, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\n`TorchScriptTensor(f32[1, 1024, 1024])`,\n)\n- onnx_kwargs: Dict[length=1](\nalpha: 1,\n)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_logical_or)`\n### Failed: attribute mismatch!\nActual {'alpha'} vs expected set()\nThe function is not a nearest match candidate.\n## Checking perfect match...\n`TracedOnnxFunction(aten_add)`\n### Failed: attribute 'alpha' type mismatch!\nActual vs\nExpected AttrType.FLOAT\nmatch score: 1\n### Exact match is not found!\nCannot find a perfect match of symbolic overload, a nearest match is found. Please check the ONNX output carefully. \n\n## Return values\n`TracedOnnxFunction(aten_add)`", + "text":"FX Node: aten.add.Tensor. \nONNX Node: aten_add[opset=pkg.onnxscript.torch_lib;is_custom=False]. \nONNX Node: aten_logical_or[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"warning", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.add.Tensor[name=add]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=308](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 828, in forward_features\n x = self._pos_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 718, in _pos_embed\n x = x + pos_embed\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.add.Tensor[name=add]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:add[name=add]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(add)[placeholder]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_dropout_Dropout)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 828, in forward_features\n x = self._pos_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 718, in _pos_embed\n x = x + pos_embed\n\n```\n## Return values\n", + "text":"FX Node: placeholder:add[name=add]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.clone.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::clone.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.clone.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.clone.default. \nONNX Node: aten_clone[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::clone.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=1](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_clone)`\nmatch score: 1\n## Return values\n`TracedOnnxFunction(aten_clone)`", + "text":"FX Node: aten.clone.default. \nONNX Node: aten_clone[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.clone.default[name=clone]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_dropout_Dropout)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\nadd: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 828, in forward_features\n x = self._pos_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 720, in _pos_embed\n return self.pos_drop(x)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.clone.default[name=clone]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: output:output[name=output]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(output)[output]:None\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_dropout_Dropout)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\nadd: `TorchScriptTensor(f32[4, 1024, 1024])`,\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: output:output[name=output]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: torch_nn_modules_dropout_Dropout. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_dropout_Dropout)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Return values\n", + "text":"FX Graph: torch_nn_modules_dropout_Dropout. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:trunk_pos_drop_1[name=trunk_pos_drop_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(trunk_pos_drop_1)[call_module]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=309](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 828, in forward_features\n x = self._pos_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 720, in _pos_embed\n return self.pos_drop(x)\n\n```\n## Return values\n", + "text":"FX Node: call_module:trunk_pos_drop_1[name=trunk_pos_drop_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:clone[name=clone]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(clone)[placeholder]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 828, in forward_features\n x = self._pos_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 720, in _pos_embed\n return self.pos_drop(x)\n\n```\n## Return values\n", + "text":"FX Node: placeholder:clone[name=clone]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_weight[name=p_getattr_l__self___trunk_blocks___0___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_weight[name=p_getattr_l__self___trunk_blocks___0___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_bias[name=p_getattr_l__self___trunk_blocks___0___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_bias[name=p_getattr_l__self___trunk_blocks___0___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___0___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=5](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___0___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___0___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=6](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___0___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_weight[name=p_getattr_l__self___trunk_blocks___0___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=7](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_weight[name=p_getattr_l__self___trunk_blocks___0___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_bias[name=p_getattr_l__self___trunk_blocks___0___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=8](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_bias[name=p_getattr_l__self___trunk_blocks___0___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=9](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=10](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=11](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=12](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm1_weight[name=p_getattr_l__self___trunk_blocks___1___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=13](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm1_weight[name=p_getattr_l__self___trunk_blocks___1___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm1_bias[name=p_getattr_l__self___trunk_blocks___1___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=14](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm1_bias[name=p_getattr_l__self___trunk_blocks___1___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___1___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=15](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___1___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___1___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=16](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___1___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___1___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=17](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___1___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___1___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=18](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___1___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm2_weight[name=p_getattr_l__self___trunk_blocks___1___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=19](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm2_weight[name=p_getattr_l__self___trunk_blocks___1___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm2_bias[name=p_getattr_l__self___trunk_blocks___1___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=20](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___norm2_bias[name=p_getattr_l__self___trunk_blocks___1___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=21](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___1___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=22](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___1___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=23](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___1___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=24](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___1___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm1_weight[name=p_getattr_l__self___trunk_blocks___2___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=25](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm1_weight[name=p_getattr_l__self___trunk_blocks___2___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm1_bias[name=p_getattr_l__self___trunk_blocks___2___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=26](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm1_bias[name=p_getattr_l__self___trunk_blocks___2___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___2___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=27](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___2___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___2___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=28](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___2___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___2___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=29](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___2___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___2___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=30](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___2___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm2_weight[name=p_getattr_l__self___trunk_blocks___2___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=31](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm2_weight[name=p_getattr_l__self___trunk_blocks___2___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm2_bias[name=p_getattr_l__self___trunk_blocks___2___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=32](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___norm2_bias[name=p_getattr_l__self___trunk_blocks___2___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=33](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___2___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=34](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___2___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=35](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___2___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=36](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___2___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm1_weight[name=p_getattr_l__self___trunk_blocks___3___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=37](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm1_weight[name=p_getattr_l__self___trunk_blocks___3___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm1_bias[name=p_getattr_l__self___trunk_blocks___3___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=38](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm1_bias[name=p_getattr_l__self___trunk_blocks___3___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___3___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=39](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___3___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___3___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=40](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___3___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___3___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=41](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___3___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___3___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=42](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___3___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm2_weight[name=p_getattr_l__self___trunk_blocks___3___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=43](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm2_weight[name=p_getattr_l__self___trunk_blocks___3___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm2_bias[name=p_getattr_l__self___trunk_blocks___3___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=44](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___norm2_bias[name=p_getattr_l__self___trunk_blocks___3___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=45](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___3___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=46](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___3___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=47](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___3___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=48](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___3___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm1_weight[name=p_getattr_l__self___trunk_blocks___4___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=49](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm1_weight[name=p_getattr_l__self___trunk_blocks___4___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm1_bias[name=p_getattr_l__self___trunk_blocks___4___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=50](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm1_bias[name=p_getattr_l__self___trunk_blocks___4___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___4___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=51](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___4___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___4___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=52](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___4___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___4___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=53](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___4___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___4___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=54](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___4___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm2_weight[name=p_getattr_l__self___trunk_blocks___4___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=55](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm2_weight[name=p_getattr_l__self___trunk_blocks___4___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm2_bias[name=p_getattr_l__self___trunk_blocks___4___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=56](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___norm2_bias[name=p_getattr_l__self___trunk_blocks___4___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=57](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___4___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=58](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___4___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=59](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___4___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=60](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___4___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm1_weight[name=p_getattr_l__self___trunk_blocks___5___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=61](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm1_weight[name=p_getattr_l__self___trunk_blocks___5___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm1_bias[name=p_getattr_l__self___trunk_blocks___5___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=62](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm1_bias[name=p_getattr_l__self___trunk_blocks___5___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___5___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=63](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___5___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___5___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=64](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___5___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___5___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=65](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___5___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___5___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=66](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___5___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm2_weight[name=p_getattr_l__self___trunk_blocks___5___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=67](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm2_weight[name=p_getattr_l__self___trunk_blocks___5___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm2_bias[name=p_getattr_l__self___trunk_blocks___5___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=68](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___norm2_bias[name=p_getattr_l__self___trunk_blocks___5___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=69](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___5___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=70](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___5___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=71](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___5___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=72](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___5___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm1_weight[name=p_getattr_l__self___trunk_blocks___6___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=73](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm1_weight[name=p_getattr_l__self___trunk_blocks___6___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm1_bias[name=p_getattr_l__self___trunk_blocks___6___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=74](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm1_bias[name=p_getattr_l__self___trunk_blocks___6___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___6___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=75](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___6___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___6___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=76](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___6___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___6___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=77](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___6___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___6___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=78](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___6___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm2_weight[name=p_getattr_l__self___trunk_blocks___6___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=79](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm2_weight[name=p_getattr_l__self___trunk_blocks___6___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm2_bias[name=p_getattr_l__self___trunk_blocks___6___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=80](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___norm2_bias[name=p_getattr_l__self___trunk_blocks___6___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=81](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___6___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=82](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___6___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=83](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___6___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=84](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___6___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm1_weight[name=p_getattr_l__self___trunk_blocks___7___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=85](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm1_weight[name=p_getattr_l__self___trunk_blocks___7___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm1_bias[name=p_getattr_l__self___trunk_blocks___7___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=86](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm1_bias[name=p_getattr_l__self___trunk_blocks___7___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___7___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=87](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___7___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___7___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=88](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___7___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___7___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=89](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___7___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___7___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=90](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___7___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm2_weight[name=p_getattr_l__self___trunk_blocks___7___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=91](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm2_weight[name=p_getattr_l__self___trunk_blocks___7___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm2_bias[name=p_getattr_l__self___trunk_blocks___7___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=92](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___norm2_bias[name=p_getattr_l__self___trunk_blocks___7___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=93](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___7___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=94](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___7___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=95](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___7___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=96](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___7___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm1_weight[name=p_getattr_l__self___trunk_blocks___8___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=97](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm1_weight[name=p_getattr_l__self___trunk_blocks___8___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm1_bias[name=p_getattr_l__self___trunk_blocks___8___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=98](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm1_bias[name=p_getattr_l__self___trunk_blocks___8___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___8___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=99](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___8___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___8___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=100](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___8___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___8___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=101](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___8___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___8___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=102](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___8___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm2_weight[name=p_getattr_l__self___trunk_blocks___8___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=103](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm2_weight[name=p_getattr_l__self___trunk_blocks___8___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm2_bias[name=p_getattr_l__self___trunk_blocks___8___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=104](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___norm2_bias[name=p_getattr_l__self___trunk_blocks___8___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=105](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___8___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=106](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___8___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=107](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___8___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=108](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___8___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm1_weight[name=p_getattr_l__self___trunk_blocks___9___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=109](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm1_weight[name=p_getattr_l__self___trunk_blocks___9___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm1_bias[name=p_getattr_l__self___trunk_blocks___9___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=110](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm1_bias[name=p_getattr_l__self___trunk_blocks___9___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___9___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=111](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___9___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___9___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=112](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___9___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___9___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=113](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___9___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___9___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=114](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___9___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm2_weight[name=p_getattr_l__self___trunk_blocks___9___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=115](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm2_weight[name=p_getattr_l__self___trunk_blocks___9___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm2_bias[name=p_getattr_l__self___trunk_blocks___9___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=116](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___norm2_bias[name=p_getattr_l__self___trunk_blocks___9___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=117](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___9___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=118](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___9___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=119](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___9___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=120](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___9___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm1_weight[name=p_getattr_l__self___trunk_blocks___10___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=121](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm1_weight[name=p_getattr_l__self___trunk_blocks___10___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm1_bias[name=p_getattr_l__self___trunk_blocks___10___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=122](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm1_bias[name=p_getattr_l__self___trunk_blocks___10___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___10___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=123](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___10___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___10___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=124](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___10___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___10___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=125](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___10___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___10___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=126](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___10___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm2_weight[name=p_getattr_l__self___trunk_blocks___10___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=127](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm2_weight[name=p_getattr_l__self___trunk_blocks___10___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm2_bias[name=p_getattr_l__self___trunk_blocks___10___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=128](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___norm2_bias[name=p_getattr_l__self___trunk_blocks___10___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=129](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___10___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=130](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___10___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=131](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___10___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=132](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___10___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm1_weight[name=p_getattr_l__self___trunk_blocks___11___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=133](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm1_weight[name=p_getattr_l__self___trunk_blocks___11___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm1_bias[name=p_getattr_l__self___trunk_blocks___11___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=134](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm1_bias[name=p_getattr_l__self___trunk_blocks___11___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___11___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=135](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___11___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___11___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=136](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___11___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___11___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=137](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___11___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___11___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=138](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___11___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm2_weight[name=p_getattr_l__self___trunk_blocks___11___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=139](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm2_weight[name=p_getattr_l__self___trunk_blocks___11___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm2_bias[name=p_getattr_l__self___trunk_blocks___11___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=140](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___norm2_bias[name=p_getattr_l__self___trunk_blocks___11___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=141](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___11___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=142](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___11___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=143](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___11___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=144](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___11___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm1_weight[name=p_getattr_l__self___trunk_blocks___12___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=145](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm1_weight[name=p_getattr_l__self___trunk_blocks___12___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm1_bias[name=p_getattr_l__self___trunk_blocks___12___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=146](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm1_bias[name=p_getattr_l__self___trunk_blocks___12___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___12___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=147](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___12___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___12___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=148](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___12___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___12___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=149](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___12___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___12___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=150](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___12___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm2_weight[name=p_getattr_l__self___trunk_blocks___12___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=151](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm2_weight[name=p_getattr_l__self___trunk_blocks___12___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm2_bias[name=p_getattr_l__self___trunk_blocks___12___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=152](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___norm2_bias[name=p_getattr_l__self___trunk_blocks___12___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=153](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___12___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=154](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___12___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=155](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___12___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=156](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___12___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm1_weight[name=p_getattr_l__self___trunk_blocks___13___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=157](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm1_weight[name=p_getattr_l__self___trunk_blocks___13___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm1_bias[name=p_getattr_l__self___trunk_blocks___13___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=158](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm1_bias[name=p_getattr_l__self___trunk_blocks___13___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___13___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=159](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___13___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___13___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=160](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___13___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___13___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=161](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___13___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___13___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=162](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___13___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm2_weight[name=p_getattr_l__self___trunk_blocks___13___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=163](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm2_weight[name=p_getattr_l__self___trunk_blocks___13___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm2_bias[name=p_getattr_l__self___trunk_blocks___13___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=164](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___norm2_bias[name=p_getattr_l__self___trunk_blocks___13___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=165](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___13___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=166](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___13___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=167](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___13___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=168](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___13___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm1_weight[name=p_getattr_l__self___trunk_blocks___14___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=169](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm1_weight[name=p_getattr_l__self___trunk_blocks___14___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm1_bias[name=p_getattr_l__self___trunk_blocks___14___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=170](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm1_bias[name=p_getattr_l__self___trunk_blocks___14___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___14___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=171](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___14___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___14___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=172](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___14___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___14___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=173](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___14___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___14___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=174](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___14___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm2_weight[name=p_getattr_l__self___trunk_blocks___14___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=175](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm2_weight[name=p_getattr_l__self___trunk_blocks___14___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm2_bias[name=p_getattr_l__self___trunk_blocks___14___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=176](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___norm2_bias[name=p_getattr_l__self___trunk_blocks___14___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=177](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___14___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=178](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___14___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=179](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___14___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=180](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___14___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm1_weight[name=p_getattr_l__self___trunk_blocks___15___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=181](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm1_weight[name=p_getattr_l__self___trunk_blocks___15___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm1_bias[name=p_getattr_l__self___trunk_blocks___15___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=182](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm1_bias[name=p_getattr_l__self___trunk_blocks___15___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___15___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=183](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___15___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___15___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=184](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___15___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___15___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=185](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___15___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___15___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=186](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___15___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm2_weight[name=p_getattr_l__self___trunk_blocks___15___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=187](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm2_weight[name=p_getattr_l__self___trunk_blocks___15___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm2_bias[name=p_getattr_l__self___trunk_blocks___15___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=188](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___norm2_bias[name=p_getattr_l__self___trunk_blocks___15___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=189](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___15___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=190](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___15___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=191](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___15___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=192](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___15___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm1_weight[name=p_getattr_l__self___trunk_blocks___16___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=193](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm1_weight[name=p_getattr_l__self___trunk_blocks___16___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm1_bias[name=p_getattr_l__self___trunk_blocks___16___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=194](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm1_bias[name=p_getattr_l__self___trunk_blocks___16___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___16___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=195](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___16___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___16___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=196](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___16___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___16___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=197](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___16___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___16___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=198](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___16___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm2_weight[name=p_getattr_l__self___trunk_blocks___16___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=199](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm2_weight[name=p_getattr_l__self___trunk_blocks___16___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm2_bias[name=p_getattr_l__self___trunk_blocks___16___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=200](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___norm2_bias[name=p_getattr_l__self___trunk_blocks___16___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=201](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___16___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=202](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___16___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=203](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___16___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=204](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___16___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm1_weight[name=p_getattr_l__self___trunk_blocks___17___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=205](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm1_weight[name=p_getattr_l__self___trunk_blocks___17___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm1_bias[name=p_getattr_l__self___trunk_blocks___17___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=206](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm1_bias[name=p_getattr_l__self___trunk_blocks___17___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___17___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=207](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___17___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___17___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=208](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___17___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___17___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=209](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___17___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___17___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=210](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___17___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm2_weight[name=p_getattr_l__self___trunk_blocks___17___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=211](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm2_weight[name=p_getattr_l__self___trunk_blocks___17___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm2_bias[name=p_getattr_l__self___trunk_blocks___17___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=212](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___norm2_bias[name=p_getattr_l__self___trunk_blocks___17___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=213](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___17___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=214](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___17___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=215](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___17___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=216](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___17___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm1_weight[name=p_getattr_l__self___trunk_blocks___18___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=217](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm1_weight[name=p_getattr_l__self___trunk_blocks___18___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm1_bias[name=p_getattr_l__self___trunk_blocks___18___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=218](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm1_bias[name=p_getattr_l__self___trunk_blocks___18___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___18___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=219](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___18___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___18___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=220](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___18___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___18___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=221](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___18___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___18___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=222](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___18___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm2_weight[name=p_getattr_l__self___trunk_blocks___18___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=223](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm2_weight[name=p_getattr_l__self___trunk_blocks___18___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm2_bias[name=p_getattr_l__self___trunk_blocks___18___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=224](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___norm2_bias[name=p_getattr_l__self___trunk_blocks___18___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=225](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___18___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=226](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___18___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=227](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___18___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=228](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___18___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm1_weight[name=p_getattr_l__self___trunk_blocks___19___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=229](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm1_weight[name=p_getattr_l__self___trunk_blocks___19___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm1_bias[name=p_getattr_l__self___trunk_blocks___19___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=230](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm1_bias[name=p_getattr_l__self___trunk_blocks___19___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___19___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=231](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___19___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___19___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=232](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___19___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___19___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=233](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___19___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___19___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=234](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___19___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm2_weight[name=p_getattr_l__self___trunk_blocks___19___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=235](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm2_weight[name=p_getattr_l__self___trunk_blocks___19___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm2_bias[name=p_getattr_l__self___trunk_blocks___19___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=236](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___norm2_bias[name=p_getattr_l__self___trunk_blocks___19___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=237](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___19___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=238](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___19___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=239](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___19___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=240](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___19___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm1_weight[name=p_getattr_l__self___trunk_blocks___20___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=241](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm1_weight[name=p_getattr_l__self___trunk_blocks___20___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm1_bias[name=p_getattr_l__self___trunk_blocks___20___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=242](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm1_bias[name=p_getattr_l__self___trunk_blocks___20___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___20___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=243](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___20___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___20___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=244](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___20___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___20___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=245](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___20___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___20___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=246](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___20___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm2_weight[name=p_getattr_l__self___trunk_blocks___20___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=247](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm2_weight[name=p_getattr_l__self___trunk_blocks___20___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm2_bias[name=p_getattr_l__self___trunk_blocks___20___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=248](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___norm2_bias[name=p_getattr_l__self___trunk_blocks___20___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=249](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___20___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=250](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___20___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=251](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___20___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=252](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___20___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm1_weight[name=p_getattr_l__self___trunk_blocks___21___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=253](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm1_weight[name=p_getattr_l__self___trunk_blocks___21___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm1_bias[name=p_getattr_l__self___trunk_blocks___21___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=254](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm1_bias[name=p_getattr_l__self___trunk_blocks___21___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___21___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=255](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___21___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___21___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=256](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___21___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___21___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=257](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___21___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___21___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=258](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___21___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm2_weight[name=p_getattr_l__self___trunk_blocks___21___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=259](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm2_weight[name=p_getattr_l__self___trunk_blocks___21___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm2_bias[name=p_getattr_l__self___trunk_blocks___21___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=260](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___norm2_bias[name=p_getattr_l__self___trunk_blocks___21___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=261](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___21___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=262](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___21___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=263](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___21___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=264](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___21___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm1_weight[name=p_getattr_l__self___trunk_blocks___22___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=265](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm1_weight[name=p_getattr_l__self___trunk_blocks___22___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm1_bias[name=p_getattr_l__self___trunk_blocks___22___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=266](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm1_bias[name=p_getattr_l__self___trunk_blocks___22___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___22___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=267](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___22___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___22___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=268](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___22___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___22___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=269](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___22___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___22___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=270](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___22___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm2_weight[name=p_getattr_l__self___trunk_blocks___22___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=271](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm2_weight[name=p_getattr_l__self___trunk_blocks___22___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm2_bias[name=p_getattr_l__self___trunk_blocks___22___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=272](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___norm2_bias[name=p_getattr_l__self___trunk_blocks___22___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=273](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___22___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=274](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___22___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=275](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___22___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=276](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___22___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm1_weight[name=p_getattr_l__self___trunk_blocks___23___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=277](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm1_weight[name=p_getattr_l__self___trunk_blocks___23___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm1_bias[name=p_getattr_l__self___trunk_blocks___23___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=278](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm1_bias[name=p_getattr_l__self___trunk_blocks___23___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___23___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=279](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___23___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___23___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=280](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___23___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___23___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=281](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___23___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___23___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=282](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___23___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm2_weight[name=p_getattr_l__self___trunk_blocks___23___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=283](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm2_weight[name=p_getattr_l__self___trunk_blocks___23___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm2_bias[name=p_getattr_l__self___trunk_blocks___23___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=284](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___norm2_bias[name=p_getattr_l__self___trunk_blocks___23___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=285](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___23___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=286](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___23___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=287](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___23___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=288](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___23___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:clone[name=clone]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(clone)[placeholder]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 828, in forward_features\n x = self._pos_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 720, in _pos_embed\n return self.pos_drop(x)\n\n```\n## Return values\n", + "text":"FX Node: placeholder:clone[name=clone]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_weight[name=p_getattr_l__self___trunk_blocks___0___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_weight[name=p_getattr_l__self___trunk_blocks___0___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_bias[name=p_getattr_l__self___trunk_blocks___0___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_bias[name=p_getattr_l__self___trunk_blocks___0___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___0___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=5](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___0___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___0___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=6](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___0___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_weight[name=p_getattr_l__self___trunk_blocks___0___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=7](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_weight[name=p_getattr_l__self___trunk_blocks___0___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_bias[name=p_getattr_l__self___trunk_blocks___0___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=8](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_bias[name=p_getattr_l__self___trunk_blocks___0___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=9](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=10](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=11](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=12](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:clone[name=clone]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(clone)[placeholder]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 828, in forward_features\n x = self._pos_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 720, in _pos_embed\n return self.pos_drop(x)\n\n```\n## Return values\n", + "text":"FX Node: placeholder:clone[name=clone]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_weight[name=p_getattr_l__self___trunk_blocks___0___norm1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm1_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_weight[name=p_getattr_l__self___trunk_blocks___0___norm1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_bias[name=p_getattr_l__self___trunk_blocks___0___norm1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm1_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm1_bias[name=p_getattr_l__self___trunk_blocks___0___norm1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.native_layer_norm.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::native_layer_norm.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.native_layer_norm.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.native_layer_norm.default. \nONNX Node: aten_native_layer_norm[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::native_layer_norm.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=5](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\nList[length=1](\n1024,\n),\n`TorchScriptTensor(f32[1024])`,\n`TorchScriptTensor(f32[1024])`,\n1e-06,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_native_layer_norm)`\nmatch score: 2\n## Return values\n`TracedOnnxFunction(aten_native_layer_norm)`", + "text":"FX Node: aten.native_layer_norm.default. \nONNX Node: aten_native_layer_norm[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.native_layer_norm.default[name=native_layer_norm]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.native_layer_norm.default[name=native_layer_norm]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:[name=getitem]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\nnative_layer_norm: Tuple[length=3](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\n`TorchScriptTensor(f32[4, 1024, 1])`,\n`TorchScriptTensor(f32[4, 1024, 1])`,\n),\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n\n```\n## Return values\n", + "text":"FX Node: call_function:[name=getitem]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: output:output[name=output]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(output)[output]:None\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=5](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\nnative_layer_norm: Tuple[length=3](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\n`TorchScriptTensor(f32[4, 1024, 1])`,\n`TorchScriptTensor(f32[4, 1024, 1])`,\n),\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: output:output[name=output]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: torch_nn_modules_normalization_LayerNorm. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Return values\n", + "text":"FX Graph: torch_nn_modules_normalization_LayerNorm. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:getattr_L__self___trunk_blocks___0___norm1_1[name=getattr_l__self___trunk_blocks___0___norm1_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getattr_L__self___trunk_blocks___0___norm1_1)[call_module]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=13](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n\n```\n## Return values\n", + "text":"FX Node: call_module:getattr_L__self___trunk_blocks___0___norm1_1[name=getattr_l__self___trunk_blocks___0___norm1_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:getitem[name=getitem]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getitem)[placeholder]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n\n```\n## Return values\n", + "text":"FX Node: placeholder:getitem[name=getitem]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___0___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___0___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___0___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___0___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:getitem[name=getitem]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getitem)[placeholder]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n\n```\n## Return values\n", + "text":"FX Node: placeholder:getitem[name=getitem]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_weight)[placeholder]:Tensor(f32[3072, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_weight[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_qkv_bias)[placeholder]:Tensor(f32[3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_qkv_bias[name=p_getattr_l__self___trunk_blocks___0___attn_qkv_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.view.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.view.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\nList[length=2](\n4096,\n1024,\n),\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_view)`\nmatch score: 2\n## Return values\n`TracedOnnxFunction(aten_view)`", + "text":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.view.default[name=view_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 88, in forward\n qkv = self.qkv(x).reshape(B, N, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.view.default[name=view_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.t.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::t.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.t.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.t.default. \nONNX Node: aten_t[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::t.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=1](\n`TorchScriptTensor(f32[3072, 1024])`,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_t)`\nmatch score: 1\n## Return values\n`TracedOnnxFunction(aten_t)`", + "text":"FX Node: aten.t.default. \nONNX Node: aten_t[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.t.default[name=t]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\nview_1: `TorchScriptTensor(f32[4096, 1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 88, in forward\n qkv = self.qkv(x).reshape(B, N, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.t.default[name=t]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.addmm.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::addmm.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.addmm.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.addmm.default. \nONNX Node: aten_addmm[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::addmm.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=3](\n`TorchScriptTensor(f32[3072])`,\n`TorchScriptTensor(f32[4096, 1024])`,\n`TorchScriptTensor(f32[1024, 3072])`,\n)\n- onnx_kwargs: Dict[length=2](\nbeta: 1,\nalpha: 1,\n)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_addmm)`\n### Failed: attribute 'beta' type mismatch!\nActual vs\nExpected AttrType.FLOAT\n### Failed: attribute 'alpha' type mismatch!\nActual vs\nExpected AttrType.FLOAT\nmatch score: 1\n### Exact match is not found!\nCannot find a perfect match of symbolic overload, a nearest match is found. Please check the ONNX output carefully. \n\n## Return values\n`TracedOnnxFunction(aten_addmm)`", + "text":"FX Node: aten.addmm.default. \nONNX Node: aten_addmm[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"warning", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.addmm.default[name=addmm]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=5](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\nview_1: `TorchScriptTensor(f32[4096, 1024])`,\nt: `TorchScriptTensor(f32[1024, 3072])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 88, in forward\n qkv = self.qkv(x).reshape(B, N, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.addmm.default[name=addmm]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.view.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.view.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor(f32[4096, 3072])`,\nList[length=3](\n4,\n1024,\n3072,\n),\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_view)`\nmatch score: 2\n## Return values\n`TracedOnnxFunction(aten_view)`", + "text":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.view.default[name=view_2]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3072])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=6](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\nview_1: `TorchScriptTensor(f32[4096, 1024])`,\nt: `TorchScriptTensor(f32[1024, 3072])`,\naddmm: `TorchScriptTensor(f32[4096, 3072])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 88, in forward\n qkv = self.qkv(x).reshape(B, N, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.view.default[name=view_2]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: output:output[name=output]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(output)[output]:None\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=7](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\nview_1: `TorchScriptTensor(f32[4096, 1024])`,\nt: `TorchScriptTensor(f32[1024, 3072])`,\naddmm: `TorchScriptTensor(f32[4096, 3072])`,\nview_2: `TorchScriptTensor(f32[4, 1024, 3072])`,\n)\n## Return values\n", + "text":"FX Node: output:output[name=output]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: torch_nn_modules_linear_Linear. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Return values\n", + "text":"FX Graph: torch_nn_modules_linear_Linear. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:getattr_L__self___trunk_blocks___0___attn_qkv_1[name=getattr_l__self___trunk_blocks___0___attn_qkv_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getattr_L__self___trunk_blocks___0___attn_qkv_1)[call_module]:Tensor(f32[4, 1024, 3072])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=5](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 88, in forward\n qkv = self.qkv(x).reshape(B, N, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4)\n\n```\n## Return values\n", + "text":"FX Node: call_module:getattr_L__self___trunk_blocks___0___attn_qkv_1[name=getattr_l__self___trunk_blocks___0___attn_qkv_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.view.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.view.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor(f32[4, 1024, 3072])`,\nList[length=5](\n4,\n1024,\n3,\n16,\n64,\n),\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_view)`\nmatch score: 2\n## Return values\n`TracedOnnxFunction(aten_view)`", + "text":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.view.default[name=view_3]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 3, 16, 64])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=6](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 88, in forward\n qkv = self.qkv(x).reshape(B, N, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.view.default[name=view_3]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.permute.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::permute.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.permute.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.permute.default. \nONNX Node: aten_permute[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::permute.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\nList[length=5](\n2,\n0,\n3,\n1,\n4,\n),\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_permute)`\nmatch score: 0\n## Return values\n`TracedOnnxFunction(aten_permute)`", + "text":"FX Node: aten.permute.default. \nONNX Node: aten_permute[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.permute.default[name=permute]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.permute.default)[call_function]:Tensor(f32[3, 4, 16, 1024, 64])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=7](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\nview_3: `TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 88, in forward\n qkv = self.qkv(x).reshape(B, N, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.permute.default[name=permute]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.unbind.int' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::unbind.int, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.unbind.int' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.unbind.int. \nONNX Node: aten_unbind[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::unbind.int, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=1](\n`TorchScriptTensor(f32[3, 4, 16, 1024, 64])`,\n)\n- onnx_kwargs: Dict[length=1](\ndim: 0,\n)\n- diagnostic_context: \n## Checking perfect match...\n`OnnxFunction(aten_unbind)`\nmatch score: 1\n## Return values\n`OnnxFunction(aten_unbind)`", + "text":"FX Node: aten.unbind.int. \nONNX Node: aten_unbind[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.unbind.int[name=unbind]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.unbind.int)[call_function]:List[length=3](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 1024, 64]),\n)\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=8](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\nview_3: `TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\npermute: `TorchScriptTensor(f32[3, 4, 16, 1024, 64])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 89, in forward\n q, k, v = qkv.unbind(0)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.unbind.int[name=unbind]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: '' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::getitem.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: '' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: . \nONNX Node: aten_getitem[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::getitem.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor()`,\n0,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`OnnxFunction(aten_getitem)`\n### Failed: input type mismatch for input 'self'!\nActual set() vs\nExpected {'seq(tensor(int8))', 'seq(tensor(complex64))', 'seq(tensor(complex128))', 'seq(tensor(int32))', 'seq(tensor(bfloat16))', 'seq(tensor(float))', 'seq(tensor(int64))', 'seq(tensor(int16))', 'seq(tensor(bool))', 'seq(tensor(uint8))', 'seq(tensor(float16))', 'seq(tensor(double))'}\nmatch score: 1\n### Exact match is not found!\nCannot find a perfect match of symbolic overload, a nearest match is found. Please check the ONNX output carefully. \n\n## Return values\n`OnnxFunction(aten_getitem)`", + "text":"FX Node: . \nONNX Node: aten_getitem[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"warning", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:[name=getitem_3]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=9](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\nview_3: `TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\npermute: `TorchScriptTensor(f32[3, 4, 16, 1024, 64])`,\nunbind: `TorchScriptTensor()`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 89, in forward\n q, k, v = qkv.unbind(0)\n\n```\n## Return values\n", + "text":"FX Node: call_function:[name=getitem_3]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: '' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::getitem.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: '' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: . \nONNX Node: aten_getitem[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::getitem.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor()`,\n1,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`OnnxFunction(aten_getitem)`\n### Failed: input type mismatch for input 'self'!\nActual set() vs\nExpected {'seq(tensor(int8))', 'seq(tensor(complex64))', 'seq(tensor(complex128))', 'seq(tensor(int32))', 'seq(tensor(bfloat16))', 'seq(tensor(float))', 'seq(tensor(int64))', 'seq(tensor(int16))', 'seq(tensor(bool))', 'seq(tensor(uint8))', 'seq(tensor(float16))', 'seq(tensor(double))'}\nmatch score: 1\n### Exact match is not found!\nCannot find a perfect match of symbolic overload, a nearest match is found. Please check the ONNX output carefully. \n\n## Return values\n`OnnxFunction(aten_getitem)`", + "text":"FX Node: . \nONNX Node: aten_getitem[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"warning", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:[name=getitem_4]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=10](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\nview_3: `TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\npermute: `TorchScriptTensor(f32[3, 4, 16, 1024, 64])`,\nunbind: `TorchScriptTensor()`,\ngetitem_3: `TorchScriptTensor(f32[4, 16, 1024, 64])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 89, in forward\n q, k, v = qkv.unbind(0)\n\n```\n## Return values\n", + "text":"FX Node: call_function:[name=getitem_4]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: '' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::getitem.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: '' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: . \nONNX Node: aten_getitem[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::getitem.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor()`,\n2,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`OnnxFunction(aten_getitem)`\n### Failed: input type mismatch for input 'self'!\nActual set() vs\nExpected {'seq(tensor(int8))', 'seq(tensor(complex64))', 'seq(tensor(complex128))', 'seq(tensor(int32))', 'seq(tensor(bfloat16))', 'seq(tensor(float))', 'seq(tensor(int64))', 'seq(tensor(int16))', 'seq(tensor(bool))', 'seq(tensor(uint8))', 'seq(tensor(float16))', 'seq(tensor(double))'}\nmatch score: 1\n### Exact match is not found!\nCannot find a perfect match of symbolic overload, a nearest match is found. Please check the ONNX output carefully. \n\n## Return values\n`OnnxFunction(aten_getitem)`", + "text":"FX Node: . \nONNX Node: aten_getitem[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"warning", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:[name=getitem_5]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=11](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\nview_3: `TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\npermute: `TorchScriptTensor(f32[3, 4, 16, 1024, 64])`,\nunbind: `TorchScriptTensor()`,\ngetitem_3: `TorchScriptTensor(f32[4, 16, 1024, 64])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 89, in forward\n q, k, v = qkv.unbind(0)\n\n```\n## Return values\n", + "text":"FX Node: call_function:[name=getitem_5]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten._scaled_dot_product_efficient_attention.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::_scaled_dot_product_efficient_attention.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten._scaled_dot_product_efficient_attention.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten._scaled_dot_product_efficient_attention.default. \nONNX Node: aten__scaled_dot_product_efficient_attention[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::_scaled_dot_product_efficient_attention.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=5](\n`TorchScriptTensor(f32[4, 16, 1024, 64])`,\n`TorchScriptTensor(f32[4, 16, 1024, 64])`,\n`TorchScriptTensor(f32[4, 16, 1024, 64])`,\n,\nFalse,\n)\n- onnx_kwargs: Dict[length=3](\ndropout_p: 0.0,\nis_causal: False,\nscale: ,\n)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten__scaled_dot_product_efficient_attention)`\n### Failed: attribute 'scale' type mismatch!\nActual vs\nExpected AttrType.FLOAT\nmatch score: 2\n### Exact match is not found!\nCannot find a perfect match of symbolic overload, a nearest match is found. Please check the ONNX output carefully. \n\n## Return values\n`TracedOnnxFunction(aten__scaled_dot_product_efficient_attention)`", + "text":"FX Node: aten._scaled_dot_product_efficient_attention.default. \nONNX Node: aten__scaled_dot_product_efficient_attention[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"warning", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten._scaled_dot_product_efficient_attention.default[name=_scaled_dot_product_efficient_attention]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten._scaled_dot_product_efficient_attention.default)[call_function]:Tuple[length=4](\nTensor(f32[4, 16, 1024, 64]),\nTensor(f32[4, 16, 0]),\nTensor(i64[]),\nTensor(i64[]),\n)\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=12](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\nview_3: `TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\npermute: `TorchScriptTensor(f32[3, 4, 16, 1024, 64])`,\nunbind: `TorchScriptTensor()`,\ngetitem_3: `TorchScriptTensor(f32[4, 16, 1024, 64])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 93, in forward\n x = F.scaled_dot_product_attention(\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten._scaled_dot_product_efficient_attention.default[name=_scaled_dot_product_efficient_attention]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:[name=getitem_6]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 16, 1024, 64])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=13](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\nview_3: `TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\npermute: `TorchScriptTensor(f32[3, 4, 16, 1024, 64])`,\nunbind: `TorchScriptTensor()`,\ngetitem_3: `TorchScriptTensor(f32[4, 16, 1024, 64])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 93, in forward\n x = F.scaled_dot_product_attention(\n\n```\n## Return values\n", + "text":"FX Node: call_function:[name=getitem_6]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.transpose.int' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::transpose.int, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.transpose.int' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.transpose.int. \nONNX Node: aten_transpose[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::transpose.int, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=3](\n`TorchScriptTensor(f32[4, 16, 1024, 64])`,\n1,\n2,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_transpose)`\nmatch score: 1\n## Return values\n`TracedOnnxFunction(aten_transpose)`", + "text":"FX Node: aten.transpose.int. \nONNX Node: aten_transpose[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.transpose.int[name=transpose_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.transpose.int)[call_function]:Tensor(f32[4, 1024, 16, 64])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=14](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\nview_3: `TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\npermute: `TorchScriptTensor(f32[3, 4, 16, 1024, 64])`,\nunbind: `TorchScriptTensor()`,\ngetitem_3: `TorchScriptTensor(f32[4, 16, 1024, 64])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 104, in forward\n x = x.transpose(1, 2).reshape(B, N, C)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.transpose.int[name=transpose_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.view.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.view.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor(f32[4, 1024, 16, 64])`,\nList[length=3](\n4,\n1024,\n1024,\n),\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_view)`\nmatch score: 2\n## Return values\n`TracedOnnxFunction(aten_view)`", + "text":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.view.default[name=view_4]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=15](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\nview_3: `TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\npermute: `TorchScriptTensor(f32[3, 4, 16, 1024, 64])`,\nunbind: `TorchScriptTensor()`,\ngetitem_3: `TorchScriptTensor(f32[4, 16, 1024, 64])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 104, in forward\n x = x.transpose(1, 2).reshape(B, N, C)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.view.default[name=view_4]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:view_4[name=view_4]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(view_4)[placeholder]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 104, in forward\n x = x.transpose(1, 2).reshape(B, N, C)\n\n```\n## Return values\n", + "text":"FX Node: placeholder:view_4[name=view_4]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___0___attn_proj_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_weight)[placeholder]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\nview_4: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_weight[name=p_getattr_l__self___trunk_blocks___0___attn_proj_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___0___attn_proj_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___attn_proj_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\nview_4: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___attn_proj_bias[name=p_getattr_l__self___trunk_blocks___0___attn_proj_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.view.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.view.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\nList[length=2](\n4096,\n1024,\n),\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_view)`\nmatch score: 2\n## Return values\n`TracedOnnxFunction(aten_view)`", + "text":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.view.default[name=view_5]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\nview_4: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 105, in forward\n x = self.proj(x)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.view.default[name=view_5]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.t.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::t.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.t.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.t.default. \nONNX Node: aten_t[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::t.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=1](\n`TorchScriptTensor(f32[1024, 1024])`,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_t)`\nmatch score: 1\n## Return values\n`TracedOnnxFunction(aten_t)`", + "text":"FX Node: aten.t.default. \nONNX Node: aten_t[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.t.default[name=t_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\nview_4: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\nview_5: `TorchScriptTensor(f32[4096, 1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 105, in forward\n x = self.proj(x)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.t.default[name=t_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.addmm.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::addmm.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.addmm.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.addmm.default. \nONNX Node: aten_addmm[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::addmm.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=3](\n`TorchScriptTensor(f32[1024])`,\n`TorchScriptTensor(f32[4096, 1024])`,\n`TorchScriptTensor(f32[1024, 1024])`,\n)\n- onnx_kwargs: Dict[length=2](\nbeta: 1,\nalpha: 1,\n)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_addmm)`\n### Failed: attribute 'beta' type mismatch!\nActual vs\nExpected AttrType.FLOAT\n### Failed: attribute 'alpha' type mismatch!\nActual vs\nExpected AttrType.FLOAT\nmatch score: 1\n### Exact match is not found!\nCannot find a perfect match of symbolic overload, a nearest match is found. Please check the ONNX output carefully. \n\n## Return values\n`TracedOnnxFunction(aten_addmm)`", + "text":"FX Node: aten.addmm.default. \nONNX Node: aten_addmm[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"warning", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.addmm.default[name=addmm_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=5](\nview_4: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\nview_5: `TorchScriptTensor(f32[4096, 1024])`,\nt_1: `TorchScriptTensor(f32[1024, 1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 105, in forward\n x = self.proj(x)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.addmm.default[name=addmm_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.view.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.view.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor(f32[4096, 1024])`,\nList[length=3](\n4,\n1024,\n1024,\n),\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_view)`\nmatch score: 2\n## Return values\n`TracedOnnxFunction(aten_view)`", + "text":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.view.default[name=view_6]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=6](\nview_4: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\nview_5: `TorchScriptTensor(f32[4096, 1024])`,\nt_1: `TorchScriptTensor(f32[1024, 1024])`,\naddmm_1: `TorchScriptTensor(f32[4096, 1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 105, in forward\n x = self.proj(x)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.view.default[name=view_6]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: output:output[name=output]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(output)[output]:None\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=7](\nview_4: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\nview_5: `TorchScriptTensor(f32[4096, 1024])`,\nt_1: `TorchScriptTensor(f32[1024, 1024])`,\naddmm_1: `TorchScriptTensor(f32[4096, 1024])`,\nview_6: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: output:output[name=output]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: torch_nn_modules_linear_Linear. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Return values\n", + "text":"FX Graph: torch_nn_modules_linear_Linear. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:getattr_L__self___trunk_blocks___0___attn_proj_1[name=getattr_l__self___trunk_blocks___0___attn_proj_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getattr_L__self___trunk_blocks___0___attn_proj_1)[call_module]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=16](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\nview_3: `TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\npermute: `TorchScriptTensor(f32[3, 4, 16, 1024, 64])`,\nunbind: `TorchScriptTensor()`,\ngetitem_3: `TorchScriptTensor(f32[4, 16, 1024, 64])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 105, in forward\n x = self.proj(x)\n\n```\n## Return values\n", + "text":"FX Node: call_module:getattr_L__self___trunk_blocks___0___attn_proj_1[name=getattr_l__self___trunk_blocks___0___attn_proj_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:view_6[name=view_6]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(view_6)[placeholder]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_dropout_Dropout)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 105, in forward\n x = self.proj(x)\n\n```\n## Return values\n", + "text":"FX Node: placeholder:view_6[name=view_6]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.clone.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::clone.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.clone.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.clone.default. \nONNX Node: aten_clone[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::clone.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=1](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_clone)`\nmatch score: 1\n## Return values\n`TracedOnnxFunction(aten_clone)`", + "text":"FX Node: aten.clone.default. \nONNX Node: aten_clone[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.clone.default[name=clone_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.clone.default)[call_function]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_dropout_Dropout)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\nview_6: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 106, in forward\n x = self.proj_drop(x)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.clone.default[name=clone_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: output:output[name=output]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(output)[output]:None\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_dropout_Dropout)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\nview_6: `TorchScriptTensor(f32[4, 1024, 1024])`,\nclone_1: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: output:output[name=output]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: torch_nn_modules_dropout_Dropout. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_dropout_Dropout)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Return values\n", + "text":"FX Graph: torch_nn_modules_dropout_Dropout. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:getattr_L__self___trunk_blocks___0___attn_proj_drop_1[name=getattr_l__self___trunk_blocks___0___attn_proj_drop_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getattr_L__self___trunk_blocks___0___attn_proj_drop_1)[call_module]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=17](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\nview_3: `TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\npermute: `TorchScriptTensor(f32[3, 4, 16, 1024, 64])`,\nunbind: `TorchScriptTensor()`,\ngetitem_3: `TorchScriptTensor(f32[4, 16, 1024, 64])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 106, in forward\n x = self.proj_drop(x)\n\n```\n## Return values\n", + "text":"FX Node: call_module:getattr_L__self___trunk_blocks___0___attn_proj_drop_1[name=getattr_l__self___trunk_blocks___0___attn_proj_drop_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: output:output[name=output]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(output)[output]:None\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=18](\ngetitem: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___attn_qkv_1: `TorchScriptTensor(f32[4, 1024, 3072])`,\nview_3: `TorchScriptTensor(f32[4, 1024, 3, 16, 64])`,\npermute: `TorchScriptTensor(f32[3, 4, 16, 1024, 64])`,\nunbind: `TorchScriptTensor()`,\ngetitem_3: `TorchScriptTensor(f32[4, 16, 1024, 64])`,\n...\n)\n## Return values\n", + "text":"FX Node: output:output[name=output]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: timm_models_vision_transformer_Attention. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Attention)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Return values\n", + "text":"FX Graph: timm_models_vision_transformer_Attention. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:getattr_L__self___trunk_blocks___0___attn_1[name=getattr_l__self___trunk_blocks___0___attn_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getattr_L__self___trunk_blocks___0___attn_1)[call_module]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=14](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 88, in forward\n qkv = self.qkv(x).reshape(B, N, 3, self.num_heads, self.head_dim).permute(2, 0, 3, 1, 4)\n\n```\n## Return values\n", + "text":"FX Node: call_module:getattr_L__self___trunk_blocks___0___attn_1[name=getattr_l__self___trunk_blocks___0___attn_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.add.Tensor' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\n- diagnostic_context: \n## Return values\nList[length=2](\nregistration.ONNXFunction(aten::add.Tensor, is_custom=False, is_complex=False),\nregistration.ONNXFunction(aten::add.Tensor, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.add.Tensor' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.add.Tensor. \nONNX Node: aten_add[opset=pkg.onnxscript.torch_lib;is_custom=False]. \nONNX Node: aten_logical_or[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\n- default_and_custom_functions: List[length=2](\nregistration.ONNXFunction(aten::add.Tensor, is_custom=False, is_complex=False),\nregistration.ONNXFunction(aten::add.Tensor, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\n`TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n- onnx_kwargs: Dict[length=1](\nalpha: 1,\n)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_logical_or)`\n### Failed: attribute mismatch!\nActual {'alpha'} vs expected set()\nThe function is not a nearest match candidate.\n## Checking perfect match...\n`TracedOnnxFunction(aten_add)`\n### Failed: attribute 'alpha' type mismatch!\nActual vs\nExpected AttrType.FLOAT\nmatch score: 1\n### Exact match is not found!\nCannot find a perfect match of symbolic overload, a nearest match is found. Please check the ONNX output carefully. \n\n## Return values\n`TracedOnnxFunction(aten_add)`", + "text":"FX Node: aten.add.Tensor. \nONNX Node: aten_add[opset=pkg.onnxscript.torch_lib;is_custom=False]. \nONNX Node: aten_logical_or[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"warning", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.add.Tensor[name=add_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.add.Tensor)[call_function]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=15](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.add.Tensor[name=add_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:add_1[name=add_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(add_1)[placeholder]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n\n```\n## Return values\n", + "text":"FX Node: placeholder:add_1[name=add_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_weight[name=p_getattr_l__self___trunk_blocks___0___norm2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm2_weight)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\nadd_1: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_weight[name=p_getattr_l__self___trunk_blocks___0___norm2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_bias[name=p_getattr_l__self___trunk_blocks___0___norm2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___norm2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\nadd_1: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___norm2_bias[name=p_getattr_l__self___trunk_blocks___0___norm2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.native_layer_norm.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::native_layer_norm.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.native_layer_norm.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.native_layer_norm.default. \nONNX Node: aten_native_layer_norm[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::native_layer_norm.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=5](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\nList[length=1](\n1024,\n),\n`TorchScriptTensor(f32[1024])`,\n`TorchScriptTensor(f32[1024])`,\n1e-06,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_native_layer_norm)`\nmatch score: 2\n## Return values\n`TracedOnnxFunction(aten_native_layer_norm)`", + "text":"FX Node: aten.native_layer_norm.default. \nONNX Node: aten_native_layer_norm[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.native_layer_norm.default[name=native_layer_norm_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.native_layer_norm.default)[call_function]:Tuple[length=3](\nTensor(f32[4, 1024, 1024]),\nTensor(f32[4, 1024, 1]),\nTensor(f32[4, 1024, 1]),\n)\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\nadd_1: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.native_layer_norm.default[name=native_layer_norm_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:[name=getitem_10]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node()[call_function]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\nadd_1: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\nnative_layer_norm_1: Tuple[length=3](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\n`TorchScriptTensor(f32[4, 1024, 1])`,\n`TorchScriptTensor(f32[4, 1024, 1])`,\n),\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n\n```\n## Return values\n", + "text":"FX Node: call_function:[name=getitem_10]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: output:output[name=output]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(output)[output]:None\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=5](\nadd_1: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\nnative_layer_norm_1: Tuple[length=3](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\n`TorchScriptTensor(f32[4, 1024, 1])`,\n`TorchScriptTensor(f32[4, 1024, 1])`,\n),\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: output:output[name=output]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: torch_nn_modules_normalization_LayerNorm. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_normalization_LayerNorm)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Return values\n", + "text":"FX Graph: torch_nn_modules_normalization_LayerNorm. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:getattr_L__self___trunk_blocks___0___norm2_1[name=getattr_l__self___trunk_blocks___0___norm2_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getattr_L__self___trunk_blocks___0___norm2_1)[call_module]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=16](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n\n```\n## Return values\n", + "text":"FX Node: call_module:getattr_L__self___trunk_blocks___0___norm2_1[name=getattr_l__self___trunk_blocks___0___norm2_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:getitem_10[name=getitem_10]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getitem_10)[placeholder]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_mlp_Mlp)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n\n```\n## Return values\n", + "text":"FX Node: placeholder:getitem_10[name=getitem_10]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_mlp_Mlp)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_mlp_Mlp)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight)[placeholder]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_mlp_Mlp)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_bias: `TorchScriptTensor(f32[4096])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias)[placeholder]:Tensor(f32[1024])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_mlp_Mlp)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_bias: `TorchScriptTensor(f32[4096])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc2_weight: `TorchScriptTensor(f32[1024, 4096])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc2_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:getitem_10[name=getitem_10]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getitem_10)[placeholder]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n\n```\n## Return values\n", + "text":"FX Node: placeholder:getitem_10[name=getitem_10]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight)[placeholder]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_weight]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias)[placeholder]:Tensor(f32[4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=2](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n)\n## Return values\n", + "text":"FX Node: placeholder:p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias[name=p_getattr_l__self___trunk_blocks___0___mlp_fc1_bias]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.view.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.view.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor(f32[4, 1024, 1024])`,\nList[length=2](\n4096,\n1024,\n),\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_view)`\nmatch score: 2\n## Return values\n`TracedOnnxFunction(aten_view)`", + "text":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.view.default[name=view_7]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4096, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=3](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_bias: `TorchScriptTensor(f32[4096])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/mlp.py\", line 44, in forward\n x = self.fc1(x)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.view.default[name=view_7]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.t.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::t.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.t.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.t.default. \nONNX Node: aten_t[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::t.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=1](\n`TorchScriptTensor(f32[4096, 1024])`,\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_t)`\nmatch score: 1\n## Return values\n`TracedOnnxFunction(aten_t)`", + "text":"FX Node: aten.t.default. \nONNX Node: aten_t[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.t.default[name=t_2]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.t.default)[call_function]:Tensor(f32[1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=4](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_bias: `TorchScriptTensor(f32[4096])`,\nview_7: `TorchScriptTensor(f32[4096, 1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/mlp.py\", line 44, in forward\n x = self.fc1(x)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.t.default[name=t_2]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.addmm.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::addmm.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.addmm.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.addmm.default. \nONNX Node: aten_addmm[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::addmm.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=3](\n`TorchScriptTensor(f32[4096])`,\n`TorchScriptTensor(f32[4096, 1024])`,\n`TorchScriptTensor(f32[1024, 4096])`,\n)\n- onnx_kwargs: Dict[length=2](\nbeta: 1,\nalpha: 1,\n)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_addmm)`\n### Failed: attribute 'beta' type mismatch!\nActual vs\nExpected AttrType.FLOAT\n### Failed: attribute 'alpha' type mismatch!\nActual vs\nExpected AttrType.FLOAT\nmatch score: 1\n### Exact match is not found!\nCannot find a perfect match of symbolic overload, a nearest match is found. Please check the ONNX output carefully. \n\n## Return values\n`TracedOnnxFunction(aten_addmm)`", + "text":"FX Node: aten.addmm.default. \nONNX Node: aten_addmm[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"warning", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.addmm.default[name=addmm_2]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.addmm.default)[call_function]:Tensor(f32[4096, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=5](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_bias: `TorchScriptTensor(f32[4096])`,\nview_7: `TorchScriptTensor(f32[4096, 1024])`,\nt_2: `TorchScriptTensor(f32[1024, 4096])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/mlp.py\", line 44, in forward\n x = self.fc1(x)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.addmm.default[name=addmm_2]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.view.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.view.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::view.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=2](\n`TorchScriptTensor(f32[4096, 4096])`,\nList[length=3](\n4,\n1024,\n4096,\n),\n)\n- onnx_kwargs: Dict[length=0](\nNone)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_view)`\nmatch score: 2\n## Return values\n`TracedOnnxFunction(aten_view)`", + "text":"FX Node: aten.view.default. \nONNX Node: aten_view[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.view.default[name=view_8]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.view.default)[call_function]:Tensor(f32[4, 1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=6](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_bias: `TorchScriptTensor(f32[4096])`,\nview_7: `TorchScriptTensor(f32[4096, 1024])`,\nt_2: `TorchScriptTensor(f32[1024, 4096])`,\naddmm_2: `TorchScriptTensor(f32[4096, 4096])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/mlp.py\", line 44, in forward\n x = self.fc1(x)\n\n```\n## Return values\n", + "text":"FX Node: call_function:aten.view.default[name=view_8]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: output:output[name=output]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(output)[output]:None\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=7](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_bias: `TorchScriptTensor(f32[4096])`,\nview_7: `TorchScriptTensor(f32[4096, 1024])`,\nt_2: `TorchScriptTensor(f32[1024, 4096])`,\naddmm_2: `TorchScriptTensor(f32[4096, 4096])`,\nview_8: `TorchScriptTensor(f32[4, 1024, 4096])`,\n)\n## Return values\n", + "text":"FX Node: output:output[name=output]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: torch_nn_modules_linear_Linear. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_linear_Linear)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Return values\n", + "text":"FX Graph: torch_nn_modules_linear_Linear. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:getattr_L__self___trunk_blocks___0___mlp_fc1_1[name=getattr_l__self___trunk_blocks___0___mlp_fc1_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getattr_L__self___trunk_blocks___0___mlp_fc1_1)[call_module]:Tensor(f32[4, 1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_mlp_Mlp)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=5](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_bias: `TorchScriptTensor(f32[4096])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc2_weight: `TorchScriptTensor(f32[1024, 4096])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc2_bias: `TorchScriptTensor(f32[1024])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/mlp.py\", line 44, in forward\n x = self.fc1(x)\n\n```\n## Return values\n", + "text":"FX Node: call_module:getattr_L__self___trunk_blocks___0___mlp_fc1_1[name=getattr_l__self___trunk_blocks___0___mlp_fc1_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: placeholder:view_8[name=view_8]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(view_8)[placeholder]:Tensor(f32[4, 1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_activations_GELUTanh)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=0](\nNone)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/mlp.py\", line 44, in forward\n x = self.fc1(x)\n\n```\n## Return values\n", + "text":"FX Node: placeholder:view_8[name=view_8]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"Searching operator overload: 'aten.gelu.default' in onnx registry...\n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher.get_function_overloads\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\n- diagnostic_context: \n## Return values\nList[length=1](\nregistration.ONNXFunction(aten::gelu.default, is_custom=False, is_complex=False),\n)", + "text":"Searching operator overload: 'aten.gelu.default' in onnx registry...\n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher.get_function_overloads" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":353 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0016", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: aten.gelu.default. \nONNX Node: aten_gelu[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n\n\n## Additional Message:\n\n## Function Signature\n### Function Signature OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\n- default_and_custom_functions: List[length=1](\nregistration.ONNXFunction(aten::gelu.default, is_custom=False, is_complex=False),\n)\n- onnx_args: Tuple[length=1](\n`TorchScriptTensor(f32[4, 1024, 4096])`,\n)\n- onnx_kwargs: Dict[length=1](\napproximate: tanh,\n)\n- diagnostic_context: \n## Checking perfect match...\n`TracedOnnxFunction(aten_gelu)`\nmatch score: 1\n## Return values\n`TracedOnnxFunction(aten_gelu)`", + "text":"FX Node: aten.gelu.default. \nONNX Node: aten_gelu[opset=pkg.onnxscript.torch_lib;is_custom=False]. \n" + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"informational", + "level":"none", + "locations":[ + { + "message":{ + "text":"OnnxFunctionDispatcher._find_the_perfect_or_nearest_match_onnxfunction" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/onnxfunction_dispatcher.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":199 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0014", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_function:aten.gelu.default[name=gelu]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(aten.gelu.default)[call_function]:Tensor(f32[4, 1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_activations_GELUTanh)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=1](\nview_8: `TorchScriptTensor(f32[4, 1024, 4096])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/mlp.py\", line 45, in forward\n x = self.act(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/activations.py\", line 159, in forward\n return F.gelu(input, approximate='tanh')\n\n```\n## Exception log\n```\nTraceback (most recent call last):\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 471, in run_node\n self.call_function(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 703, in call_function\n ] = symbolic_fn(*onnx_args, **onnx_kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/values.py\", line 625, in __call__\n return self.func(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 474, in aten_gelu\n result = _aten_gelu_approximate_tanh(self)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 496, in _aten_gelu_approximate_tanh\n cubed = op.Pow(self, ir.tensor(3, dtype=self.dtype))\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnx_ir/_convenience/_constructors.py\", line 103, in tensor\n raise TypeError(f\"dtype must be an instance of DataType. dtype={dtype}\")\n\nTypeError: dtype must be an instance of DataType. dtype=torch.float32\n\n```", + "text":"FX Node: call_function:aten.gelu.default[name=gelu]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"error", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: timm_layers_activations_GELUTanh. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(timm_layers_activations_GELUTanh)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Exception log\n```\nTraceback (most recent call last):\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 471, in run_node\n self.call_function(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 703, in call_function\n ] = symbolic_fn(*onnx_args, **onnx_kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/values.py\", line 625, in __call__\n return self.func(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 474, in aten_gelu\n result = _aten_gelu_approximate_tanh(self)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 496, in _aten_gelu_approximate_tanh\n cubed = op.Pow(self, ir.tensor(3, dtype=self.dtype))\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnx_ir/_convenience/_constructors.py\", line 103, in tensor\n raise TypeError(f\"dtype must be an instance of DataType. dtype={dtype}\")\n\nTypeError: dtype must be an instance of DataType. dtype=torch.float32\n\n```", + "text":"FX Graph: timm_layers_activations_GELUTanh. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"error", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:getattr_L__self___trunk_blocks___0___mlp_act_1[name=getattr_l__self___trunk_blocks___0___mlp_act_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getattr_L__self___trunk_blocks___0___mlp_act_1)[call_module]:Tensor(f32[4, 1024, 4096])\n- fx_graph_module: torch.fx.GraphModule(timm_layers_mlp_Mlp)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=6](\ngetitem_10: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_bias: `TorchScriptTensor(f32[4096])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc2_weight: `TorchScriptTensor(f32[1024, 4096])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc2_bias: `TorchScriptTensor(f32[1024])`,\ngetattr_l__self___trunk_blocks___0___mlp_fc1_1: `TorchScriptTensor(f32[4, 1024, 4096])`,\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/mlp.py\", line 45, in forward\n x = self.act(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/activations.py\", line 159, in forward\n return F.gelu(input, approximate='tanh')\n\n```\n## Exception log\n```\nTraceback (most recent call last):\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 471, in run_node\n self.call_function(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 703, in call_function\n ] = symbolic_fn(*onnx_args, **onnx_kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/values.py\", line 625, in __call__\n return self.func(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 474, in aten_gelu\n result = _aten_gelu_approximate_tanh(self)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 496, in _aten_gelu_approximate_tanh\n cubed = op.Pow(self, ir.tensor(3, dtype=self.dtype))\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnx_ir/_convenience/_constructors.py\", line 103, in tensor\n raise TypeError(f\"dtype must be an instance of DataType. dtype={dtype}\")\n\nTypeError: dtype must be an instance of DataType. dtype=torch.float32\n\n```", + "text":"FX Node: call_module:getattr_L__self___trunk_blocks___0___mlp_act_1[name=getattr_l__self___trunk_blocks___0___mlp_act_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"error", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: timm_layers_mlp_Mlp. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(timm_layers_mlp_Mlp)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Exception log\n```\nTraceback (most recent call last):\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 471, in run_node\n self.call_function(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 703, in call_function\n ] = symbolic_fn(*onnx_args, **onnx_kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/values.py\", line 625, in __call__\n return self.func(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 474, in aten_gelu\n result = _aten_gelu_approximate_tanh(self)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 496, in _aten_gelu_approximate_tanh\n cubed = op.Pow(self, ir.tensor(3, dtype=self.dtype))\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnx_ir/_convenience/_constructors.py\", line 103, in tensor\n raise TypeError(f\"dtype must be an instance of DataType. dtype={dtype}\")\n\nTypeError: dtype must be an instance of DataType. dtype=torch.float32\n\n```", + "text":"FX Graph: timm_layers_mlp_Mlp. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"error", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:getattr_L__self___trunk_blocks___0___mlp_1[name=getattr_l__self___trunk_blocks___0___mlp_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(getattr_L__self___trunk_blocks___0___mlp_1)[call_module]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=17](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 170, in forward\n x = x + self.drop_path2(self.ls2(self.mlp(self.norm2(x))))\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/mlp.py\", line 44, in forward\n x = self.fc1(x)\n\n```\n## Exception log\n```\nTraceback (most recent call last):\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 471, in run_node\n self.call_function(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 703, in call_function\n ] = symbolic_fn(*onnx_args, **onnx_kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/values.py\", line 625, in __call__\n return self.func(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 474, in aten_gelu\n result = _aten_gelu_approximate_tanh(self)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 496, in _aten_gelu_approximate_tanh\n cubed = op.Pow(self, ir.tensor(3, dtype=self.dtype))\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnx_ir/_convenience/_constructors.py\", line 103, in tensor\n raise TypeError(f\"dtype must be an instance of DataType. dtype={dtype}\")\n\nTypeError: dtype must be an instance of DataType. dtype=torch.float32\n\n```", + "text":"FX Node: call_module:getattr_L__self___trunk_blocks___0___mlp_1[name=getattr_l__self___trunk_blocks___0___mlp_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"error", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: timm_models_vision_transformer_Block. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_Block)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Exception log\n```\nTraceback (most recent call last):\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 471, in run_node\n self.call_function(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 703, in call_function\n ] = symbolic_fn(*onnx_args, **onnx_kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/values.py\", line 625, in __call__\n return self.func(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 474, in aten_gelu\n result = _aten_gelu_approximate_tanh(self)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 496, in _aten_gelu_approximate_tanh\n cubed = op.Pow(self, ir.tensor(3, dtype=self.dtype))\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnx_ir/_convenience/_constructors.py\", line 103, in tensor\n raise TypeError(f\"dtype must be an instance of DataType. dtype={dtype}\")\n\nTypeError: dtype must be an instance of DataType. dtype=torch.float32\n\n```", + "text":"FX Graph: timm_models_vision_transformer_Block. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"error", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:trunk_blocks_0_1[name=trunk_blocks_0_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(trunk_blocks_0_1)[call_module]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=289](\nclone: `TorchScriptTensor(f32[4, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm2_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___mlp_fc1_weight: `TorchScriptTensor(f32[4096, 1024])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n\n```\n## Exception log\n```\nTraceback (most recent call last):\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 471, in run_node\n self.call_function(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 703, in call_function\n ] = symbolic_fn(*onnx_args, **onnx_kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/values.py\", line 625, in __call__\n return self.func(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 474, in aten_gelu\n result = _aten_gelu_approximate_tanh(self)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 496, in _aten_gelu_approximate_tanh\n cubed = op.Pow(self, ir.tensor(3, dtype=self.dtype))\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnx_ir/_convenience/_constructors.py\", line 103, in tensor\n raise TypeError(f\"dtype must be an instance of DataType. dtype={dtype}\")\n\nTypeError: dtype must be an instance of DataType. dtype=torch.float32\n\n```", + "text":"FX Node: call_module:trunk_blocks_0_1[name=trunk_blocks_0_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"error", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: torch_nn_modules_container_Sequential. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(torch_nn_modules_container_Sequential)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Exception log\n```\nTraceback (most recent call last):\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 471, in run_node\n self.call_function(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 703, in call_function\n ] = symbolic_fn(*onnx_args, **onnx_kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/values.py\", line 625, in __call__\n return self.func(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 474, in aten_gelu\n result = _aten_gelu_approximate_tanh(self)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 496, in _aten_gelu_approximate_tanh\n cubed = op.Pow(self, ir.tensor(3, dtype=self.dtype))\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnx_ir/_convenience/_constructors.py\", line 103, in tensor\n raise TypeError(f\"dtype must be an instance of DataType. dtype={dtype}\")\n\nTypeError: dtype must be an instance of DataType. dtype=torch.float32\n\n```", + "text":"FX Graph: torch_nn_modules_container_Sequential. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"error", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:trunk_blocks_1[name=trunk_blocks_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(trunk_blocks_1)[call_module]:Tensor(f32[4, 1024, 1024])\n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=310](\nx: `TorchScriptTensor(f32[4, 3, 512, 512])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 834, in forward_features\n x = self.blocks(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 169, in forward\n x = x + self.drop_path1(self.ls1(self.attn(self.norm1(x))))\n\n```\n## Exception log\n```\nTraceback (most recent call last):\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 471, in run_node\n self.call_function(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 703, in call_function\n ] = symbolic_fn(*onnx_args, **onnx_kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/values.py\", line 625, in __call__\n return self.func(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 474, in aten_gelu\n result = _aten_gelu_approximate_tanh(self)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 496, in _aten_gelu_approximate_tanh\n cubed = op.Pow(self, ir.tensor(3, dtype=self.dtype))\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnx_ir/_convenience/_constructors.py\", line 103, in tensor\n raise TypeError(f\"dtype must be an instance of DataType. dtype={dtype}\")\n\nTypeError: dtype must be an instance of DataType. dtype=torch.float32\n\n```", + "text":"FX Node: call_module:trunk_blocks_1[name=trunk_blocks_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"error", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: timm_models_vision_transformer_VisionTransformer. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule(timm_models_vision_transformer_VisionTransformer)\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- parent_onnxscript_graph: \n## Exception log\n```\nTraceback (most recent call last):\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 471, in run_node\n self.call_function(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 703, in call_function\n ] = symbolic_fn(*onnx_args, **onnx_kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/values.py\", line 625, in __call__\n return self.func(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 474, in aten_gelu\n result = _aten_gelu_approximate_tanh(self)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 496, in _aten_gelu_approximate_tanh\n cubed = op.Pow(self, ir.tensor(3, dtype=self.dtype))\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnx_ir/_convenience/_constructors.py\", line 103, in tensor\n raise TypeError(f\"dtype must be an instance of DataType. dtype={dtype}\")\n\nTypeError: dtype must be an instance of DataType. dtype=torch.float32\n\n```", + "text":"FX Graph: timm_models_vision_transformer_VisionTransformer. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"error", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Node: call_module:trunk_1[name=trunk_1]. \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run_node\n- self: \n- node: fx.Node(trunk_1)[call_module]:Tensor(f32[4, 1024])\n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n- onnxscript_graph: \n- onnxscript_tracer: \n- fx_name_to_onnxscript_value: Dict[length=307](\np_trunk_pos_embed: `TorchScriptTensor(f32[1, 1024, 1024])`,\np_trunk_attn_pool_latent: `TorchScriptTensor(f32[1, 1, 1024])`,\np_trunk_patch_embed_proj_weight: `TorchScriptTensor(f32[1024, 3, 16, 16])`,\np_trunk_patch_embed_proj_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_weight: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___norm1_bias: `TorchScriptTensor(f32[1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_weight: `TorchScriptTensor(f32[3072, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_qkv_bias: `TorchScriptTensor(f32[3072])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_weight: `TorchScriptTensor(f32[1024, 1024])`,\np_getattr_l__self___trunk_blocks___0___attn_proj_bias: `TorchScriptTensor(f32[1024])`,\n...\n)\n## PyTorch source information\n```\n File \"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py\", line 196, in forward\n x = self.trunk(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 853, in forward\n x = self.forward_features(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/models/vision_transformer.py\", line 827, in forward_features\n x = self.patch_embed(x)\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/nn/modules/module.py\", line 1725, in _call_impl\n return forward_call(*args, **kwargs)\n File \"/home/thebears/.local/lib/python3.10/site-packages/timm/layers/patch_embed.py\", line 131, in forward\n x = self.proj(x)\n\n```\n## Exception log\n```\nTraceback (most recent call last):\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 471, in run_node\n self.call_function(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 703, in call_function\n ] = symbolic_fn(*onnx_args, **onnx_kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/values.py\", line 625, in __call__\n return self.func(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 474, in aten_gelu\n result = _aten_gelu_approximate_tanh(self)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 496, in _aten_gelu_approximate_tanh\n cubed = op.Pow(self, ir.tensor(3, dtype=self.dtype))\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnx_ir/_convenience/_constructors.py\", line 103, in tensor\n raise TypeError(f\"dtype must be an instance of DataType. dtype={dtype}\")\n\nTypeError: dtype must be an instance of DataType. dtype=torch.float32\n\n```", + "text":"FX Node: call_module:trunk_1[name=trunk_1]. " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"error", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run_node" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":414 + } + } + }, + { + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/open_clip/timm_model.py" + }, + "region":{ + "snippet":{ + "text":"x = self.trunk(x)" + }, + "startLine":196 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0008", + "stacks":[] + }, + { + "message":{ + "markdown":"FX Graph: . \n\n## Additional Message:\n\n## Function Signature\n### Function Signature FxOnnxInterpreter.run\n- self: \n- fx_graph_module: torch.fx.GraphModule()\n- onnxfunction_dispatcher: \n- op_level_debug: False\n## Exception log\n```\nTraceback (most recent call last):\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 482, in run_node\n self.call_module(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 811, in call_module\n sub_onnxscript_graph = self.run(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 577, in run\n self.run_node(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 152, in wrapper\n ctx.log_and_raise_if_error(diag)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/context.py\", line 369, in log_and_raise_if_error\n raise diagnostic.source_exception\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/diagnostics/infra/decorator.py\", line 136, in wrapper\n return_values = fn(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 471, in run_node\n self.call_function(\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py\", line 703, in call_function\n ] = symbolic_fn(*onnx_args, **onnx_kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/values.py\", line 625, in __call__\n return self.func(*args, **kwargs)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 474, in aten_gelu\n result = _aten_gelu_approximate_tanh(self)\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnxscript/function_libs/torch_lib/ops/nn.py\", line 496, in _aten_gelu_approximate_tanh\n cubed = op.Pow(self, ir.tensor(3, dtype=self.dtype))\n\n File \"/home/thebears/.local/lib/python3.10/site-packages/onnx_ir/_convenience/_constructors.py\", line 103, in tensor\n raise TypeError(f\"dtype must be an instance of DataType. dtype={dtype}\")\n\nTypeError: dtype must be an instance of DataType. dtype=torch.float32\n\n```", + "text":"FX Graph: . " + }, + "codeFlows":[ + { + "threadFlows":[ + { + "locations":[] + } + ] + } + ], + "graphs":[], + "kind":"fail", + "level":"error", + "locations":[ + { + "message":{ + "text":"FxOnnxInterpreter.run" + }, + "physicalLocation":{ + "artifactLocation":{ + "uri":"/home/thebears/.local/lib/python3.10/site-packages/torch/onnx/_internal/fx/fx_onnx_interpreter.py" + }, + "region":{ + "snippet":{ + "text":"@_beartype.beartype" + }, + "startLine":496 + } + } + } + ], + "properties":{ + "tags":[] + }, + "ruleId":"FXE0007", + "stacks":[] + } + ] + } + ], + "version":"2.1.0", + "schemaUri":"https://docs.oasis-open.org/sarif/sarif/v2.1.0/cs01/schemas/sarif-schema-2.1.0.json" +} \ No newline at end of file diff --git a/saved.png b/saved.png new file mode 100644 index 0000000..61a4d9e Binary files /dev/null and b/saved.png differ diff --git a/short.npz b/short.npz new file mode 100755 index 0000000..f8780bf Binary files /dev/null and b/short.npz differ diff --git a/try_decode.py b/try_decode.py index d43f49c..910fb4a 100644 --- a/try_decode.py +++ b/try_decode.py @@ -1,4 +1,5 @@ import time +from datetime import datetime import cv2 import numpy @@ -12,7 +13,7 @@ import torch from cuda import cuda as ccuda from cuda import cudart -cmd = "filesrc location=/home/thebears/local/source/full.mp4 ! qtdemux name=demux demux.video_0 ! queue ! h265parse ! nvv4l2decoder ! nvvidconv ! videoscale method=1 add-borders=false ! video/x-raw,width=1280,height=1280 ! appsink sync=false" +cmd = "filesrc location=/home/thebears/local/source/short.mp4 ! qtdemux name=demux demux.video_0 ! queue ! h265parse ! nvv4l2decoder ! nvvidconv ! videoscale method=1 add-borders=false ! video/x-raw,width=1280,height=1280 ! appsink sync=false" cap = cv2.VideoCapture(cmd, cv2.CAP_GSTREAMER) @@ -21,6 +22,7 @@ fr = 0 arrays_to_score = list() +imgs = list() array = list() while True: good, frf = cap.read() @@ -31,11 +33,11 @@ while True: array.append(frf) - + imgs.append(frf) if len(array) > 8: arrays_to_score.append(torch.from_numpy(np.asarray(array))) array = list() - break + if len(array) > 0: @@ -45,55 +47,84 @@ if len(array) > 0: et = time.time() print(et - st, fr / (st - et)) +# %% +from datetime import datetime +pretrained_name = "webli" +#model_name = "ViT-L-16-SigLIP2-512" +model_name = 'ViT-SO400M-16-SigLIP2-512' +rt_dir ='/home/thebears/local/source/models/' - +os.makedirs(rt_dir, exist_ok=True) +fname = model_name.replace('-','_').lower() + '_'+datetime.now().strftime('%Y%m%d') +ONNX_FILE_PATH=os.path.join(rt_dir, fname + '.onnx') +ENGINE_FILE_PATH = os.path.splitext(ONNX_FILE_PATH)[0]+'.engine' # %% -pretrained_name = "webli" -model_name = "ViT-L-16-SigLIP-512" - - -model_name = 'ViT-SO400M-16-SigLIP2-512' -ONNX_FILE_PATH = "/home/thebears/local/source/so400m_siglip2_512.onnx" -#model_name, pretrained_name = ('ViT-B-16-quickgelu', 'openai') model, _, preprocess = open_clip.create_model_and_transforms( model_name, pretrained=pretrained_name ) +# %% + +model_gpu = model.cuda() +scores = list() +all_means = list() +with torch.no_grad(): + for fr_num, img in enumerate(imgs): + + tensor_raw = torch.tensor(img[None,:,:,0:3]) + tensor_perm = tensor_raw.permute([0, 3, 1, 2]).to(torch.float32) / 255 + tensor_reshaped = preprocess.transforms[0](tensor_perm) + tensor_mean = preprocess.transforms[-1](tensor_reshaped) + all_means.append(tensor_mean) + imp = model_gpu.encode_image(tensor_mean.cuda()) + + print(fr_num) + scores.append((fr_num, imp.detach().cpu().numpy())) +# %% +np.save('dump_so400m',np.concatenate([x[1] for x in scores])) + + # %% with torch.no_grad(): et = time.time() + if True: - tensor_raw = arrays_to_score[0][0,:,:,0:3][None,:,:,:] + + tensor_raw = torch.concat(arrays_to_score)[0:4, :, :, 0:3] tensor_perm = tensor_raw.permute([0, 3, 1, 2]).to(torch.float32) / 255 tensor_reshaped = preprocess.transforms[0](tensor_perm) tensor_mean = preprocess.transforms[-1](tensor_reshaped) else: - tensor_raw = torch.concat(arrays_to_score)[0:4, :, :, 0:3] + tensor_raw = torch.concat(arrays_to_score)[0, :, :, 0:3] tensor_perm = tensor_raw.permute([0, 3, 1, 2]).to(torch.float32) / 255 tensor_reshaped = preprocess.transforms[1](preprocess.transforms[0](tensor_perm)) tensor_mean = preprocess.transforms[-1](tensor_reshaped) - imp = model.encode_image(tensor_mean) + #imp = model.encode_image(tensor_mean) + imp = model_gpu.encode_image(tensor_mean.cuda()) st = time.time() print((st - et) / tensor_raw.shape[0], tensor_raw.shape[0]/(st - et) ) + +from_model_on_gpu = imp.detach().cpu().numpy() + + - from_model_on_gpu = imp.cpu().numpy() # %% -ENGINE_FILE_PATH = os.path.splitext(ONNX_FILE_PATH)[0]+'.trt' + torch.onnx.export( - model.visual, - tensor_mean, + model.visual.cuda(), + tensor_mean.cuda(), ONNX_FILE_PATH, input_names=["input"], output_names=["output"], ) -# %% + X_test = tensor_mean.cpu().numpy() sess = rt.InferenceSession( ONNX_FILE_PATH, providers=rt.get_available_providers()) @@ -106,7 +137,7 @@ def norm(v): print(np.dot(norm(pred_onx), norm(from_model_on_gpu).T)) -# %% + TRT_LOGGER = trt.Logger() def build_engine_from_onnx(onnx_file_path, use_fp16=True): """ @@ -142,7 +173,7 @@ def build_engine_from_onnx(onnx_file_path, use_fp16=True): # Enable FP16 precision if requested and if the GPU supports it if use_fp16: if builder.platform_has_fast_fp16: -# config.set_flag(trt.BuilderFlag.FP16) + config.set_flag(trt.BuilderFlag.FP16) print("FP16 enabled successfully") else: print("Warning: GPU doesn't support fast FP16, using FP32 instead") @@ -160,7 +191,7 @@ def build_engine_from_onnx(onnx_file_path, use_fp16=True): -engine = build_engine_from_onnx(ONNX_FILE_PATH, use_fp16=False) +engine = build_engine_from_onnx(ONNX_FILE_PATH, use_fp16=True) with open(ENGINE_FILE_PATH, "wb") as f: f.write(engine)