Searched refs:split_output (Results 1 – 5 of 5) sorted by relevance
/aosp_15_r20/external/pytorch/test/inductor/ |
H A D | test_split_cat_fx_passes.py | 314 split_output = list(torch.split(x, 4, dim=1)) 316 [torch.ones(2, 5, 32, 16)] + split_output + [torch.ones(2, 6, 32, 16)], 321 split_output = list(torch.split(x, 4, dim=1)) 324 + split_output 330 split_output = list(torch.split(x, 4, dim=2)) 332 [torch.ones(2, 32, 5, 16)] + split_output + [torch.ones(2, 32, 6, 16)], 338 split_output = list(torch.split(x, 4, dim=1)) 340 [torch.ones(2, 4, 32, 16)] + split_output + [torch.ones(2, 4, 32, 16)], 345 split_output = list(torch.split(x, 4, dim=1)) 347 [torch.ones(2, 4, 32, 16)] + split_output + [torch.ones(2, 4, 32, 16)], [all …]
|
/aosp_15_r20/build/make/tools/warn/ |
H A D | warn_common.py | 514 split_output = compiler_output.split(':', 3) # 3 = max splits 515 file_path = split_output[0] 516 line_number = int(split_output[1]) 517 col_number = int(split_output[2].split(' ')[0]) 518 warning_message = split_output[3]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cuda/ |
H A D | Reduce.cuh | 118 int split_output(int parallelism) { in split_output() function 1091 config.output_mult[0] = config.split_output(block_width); in setReduceConfig() 1104 config.output_mult[1] = config.split_output(block_height); in setReduceConfig()
|
/aosp_15_r20/external/tensorflow/tensorflow/lite/toco/ |
H A D | export_tensorflow.cc | 1571 std::string split_output = base + "split"; in ConvertLstmCellOperator() local 1574 split_op->set_name(split_output); in ConvertLstmCellOperator() 1585 *tanh_0_op->add_input() = split_output + ":1"; in ConvertLstmCellOperator() 1592 *logistic_1_op->add_input() = split_output; in ConvertLstmCellOperator() 1607 *logistic_2_op->add_input() = split_output + ":2"; in ConvertLstmCellOperator() 1614 *logistic_3_op->add_input() = split_output + ":3"; in ConvertLstmCellOperator()
|
/aosp_15_r20/out/.module_paths/ |
D | files.db | 9 …8184241,"I":6295928,"F":[]},{"P":"fonttools/Tests/designspaceLib/data/split_output","T":1745562318…
|